var/home/core/zuul-output/0000755000175000017500000000000015070107551014526 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015070113677015500 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000003647371015070113670017710 0ustar rootrootOct 04 03:10:32 crc systemd[1]: Starting Kubernetes Kubelet... Oct 04 03:10:32 crc restorecon[4676]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 04 03:10:32 crc restorecon[4676]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 03:10:32 crc restorecon[4676]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 03:10:32 crc restorecon[4676]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 03:10:32 crc restorecon[4676]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 03:10:32 crc restorecon[4676]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 03:10:32 crc restorecon[4676]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 03:10:32 crc restorecon[4676]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 03:10:32 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 04 03:10:32 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 03:10:32 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 04 03:10:32 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 04 03:10:32 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 03:10:32 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 03:10:32 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 03:10:32 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 03:10:32 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 04 03:10:32 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 04 03:10:32 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:10:32 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:10:32 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:10:32 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:10:32 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 03:10:33 crc restorecon[4676]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 03:10:33 crc restorecon[4676]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 04 03:10:34 crc kubenswrapper[4742]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 04 03:10:34 crc kubenswrapper[4742]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 04 03:10:34 crc kubenswrapper[4742]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 04 03:10:34 crc kubenswrapper[4742]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 04 03:10:34 crc kubenswrapper[4742]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 04 03:10:34 crc kubenswrapper[4742]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.551661 4742 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.561932 4742 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.562160 4742 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.562241 4742 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.562361 4742 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.562447 4742 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.562526 4742 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.562643 4742 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.562730 4742 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.562870 4742 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.563070 4742 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.563161 4742 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.563252 4742 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.563361 4742 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.563439 4742 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.563573 4742 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.563671 4742 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.563805 4742 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.563890 4742 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.564027 4742 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.564117 4742 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.564204 4742 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.564322 4742 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.564411 4742 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.564508 4742 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.564596 4742 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.564702 4742 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.564795 4742 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.564885 4742 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.564965 4742 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.565031 4742 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.565092 4742 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.565152 4742 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.565266 4742 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.565430 4742 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.565507 4742 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.565572 4742 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.565699 4742 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.565769 4742 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.565830 4742 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.566000 4742 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.566077 4742 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.566138 4742 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.566199 4742 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.566263 4742 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.566352 4742 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.566415 4742 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.566516 4742 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.566594 4742 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.566661 4742 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.566723 4742 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.566787 4742 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.566849 4742 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.566968 4742 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.567110 4742 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.567182 4742 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.567258 4742 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.567350 4742 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.567413 4742 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.567486 4742 feature_gate.go:330] unrecognized feature gate: Example Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.567599 4742 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.567769 4742 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.567837 4742 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.567910 4742 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.568019 4742 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.568109 4742 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.568173 4742 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.568235 4742 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.568512 4742 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.568578 4742 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.568638 4742 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.568698 4742 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.570010 4742 flags.go:64] FLAG: --address="0.0.0.0" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.570118 4742 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.570204 4742 flags.go:64] FLAG: --anonymous-auth="true" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.570361 4742 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.570470 4742 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.570534 4742 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.570621 4742 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.570770 4742 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.570871 4742 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.570948 4742 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.571022 4742 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.571115 4742 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.571203 4742 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.571363 4742 flags.go:64] FLAG: --cgroup-root="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.571437 4742 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.571500 4742 flags.go:64] FLAG: --client-ca-file="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.571563 4742 flags.go:64] FLAG: --cloud-config="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.571639 4742 flags.go:64] FLAG: --cloud-provider="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.571725 4742 flags.go:64] FLAG: --cluster-dns="[]" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.571802 4742 flags.go:64] FLAG: --cluster-domain="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.571878 4742 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.571964 4742 flags.go:64] FLAG: --config-dir="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.572074 4742 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.572140 4742 flags.go:64] FLAG: --container-log-max-files="5" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.572222 4742 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.572430 4742 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.572586 4742 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.572676 4742 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.572753 4742 flags.go:64] FLAG: --contention-profiling="false" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.572828 4742 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.572960 4742 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.573064 4742 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.573150 4742 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.573239 4742 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.573355 4742 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.573428 4742 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.573492 4742 flags.go:64] FLAG: --enable-load-reader="false" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.573554 4742 flags.go:64] FLAG: --enable-server="true" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.573635 4742 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.573704 4742 flags.go:64] FLAG: --event-burst="100" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.573779 4742 flags.go:64] FLAG: --event-qps="50" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.573851 4742 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.573950 4742 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.574154 4742 flags.go:64] FLAG: --eviction-hard="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.574267 4742 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.574364 4742 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.574433 4742 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.574513 4742 flags.go:64] FLAG: --eviction-soft="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.574587 4742 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.574665 4742 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.574801 4742 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.574902 4742 flags.go:64] FLAG: --experimental-mounter-path="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.574982 4742 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.575102 4742 flags.go:64] FLAG: --fail-swap-on="true" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.575180 4742 flags.go:64] FLAG: --feature-gates="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.575248 4742 flags.go:64] FLAG: --file-check-frequency="20s" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.575387 4742 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.575456 4742 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.575522 4742 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.575587 4742 flags.go:64] FLAG: --healthz-port="10248" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.575652 4742 flags.go:64] FLAG: --help="false" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.575730 4742 flags.go:64] FLAG: --hostname-override="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.575803 4742 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.575872 4742 flags.go:64] FLAG: --http-check-frequency="20s" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.575951 4742 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.576044 4742 flags.go:64] FLAG: --image-credential-provider-config="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.576214 4742 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.576353 4742 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.576437 4742 flags.go:64] FLAG: --image-service-endpoint="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.576509 4742 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.576576 4742 flags.go:64] FLAG: --kube-api-burst="100" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.576689 4742 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.576757 4742 flags.go:64] FLAG: --kube-api-qps="50" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.576836 4742 flags.go:64] FLAG: --kube-reserved="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.576918 4742 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.576992 4742 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.577068 4742 flags.go:64] FLAG: --kubelet-cgroups="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.577145 4742 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.577214 4742 flags.go:64] FLAG: --lock-file="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.577298 4742 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.577381 4742 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.577475 4742 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.577555 4742 flags.go:64] FLAG: --log-json-split-stream="false" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.577653 4742 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.577747 4742 flags.go:64] FLAG: --log-text-split-stream="false" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.577828 4742 flags.go:64] FLAG: --logging-format="text" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.577917 4742 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.578065 4742 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.578206 4742 flags.go:64] FLAG: --manifest-url="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.578358 4742 flags.go:64] FLAG: --manifest-url-header="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.578473 4742 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.578565 4742 flags.go:64] FLAG: --max-open-files="1000000" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.578708 4742 flags.go:64] FLAG: --max-pods="110" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.578794 4742 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.578864 4742 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.578984 4742 flags.go:64] FLAG: --memory-manager-policy="None" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.579057 4742 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.579170 4742 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.579296 4742 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.579390 4742 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.579514 4742 flags.go:64] FLAG: --node-status-max-images="50" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.579609 4742 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.579703 4742 flags.go:64] FLAG: --oom-score-adj="-999" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.579800 4742 flags.go:64] FLAG: --pod-cidr="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.579890 4742 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.580025 4742 flags.go:64] FLAG: --pod-manifest-path="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.580121 4742 flags.go:64] FLAG: --pod-max-pids="-1" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.580202 4742 flags.go:64] FLAG: --pods-per-core="0" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.580313 4742 flags.go:64] FLAG: --port="10250" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.580411 4742 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.580555 4742 flags.go:64] FLAG: --provider-id="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.580655 4742 flags.go:64] FLAG: --qos-reserved="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.580795 4742 flags.go:64] FLAG: --read-only-port="10255" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.580883 4742 flags.go:64] FLAG: --register-node="true" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.580950 4742 flags.go:64] FLAG: --register-schedulable="true" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.581015 4742 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.581087 4742 flags.go:64] FLAG: --registry-burst="10" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.581151 4742 flags.go:64] FLAG: --registry-qps="5" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.581347 4742 flags.go:64] FLAG: --reserved-cpus="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.581462 4742 flags.go:64] FLAG: --reserved-memory="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.581538 4742 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.581604 4742 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.581728 4742 flags.go:64] FLAG: --rotate-certificates="false" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.581809 4742 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.581875 4742 flags.go:64] FLAG: --runonce="false" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.581939 4742 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.582011 4742 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.582077 4742 flags.go:64] FLAG: --seccomp-default="false" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.582140 4742 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.582203 4742 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.582267 4742 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.582368 4742 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.582445 4742 flags.go:64] FLAG: --storage-driver-password="root" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.582516 4742 flags.go:64] FLAG: --storage-driver-secure="false" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.582581 4742 flags.go:64] FLAG: --storage-driver-table="stats" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.582667 4742 flags.go:64] FLAG: --storage-driver-user="root" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.582743 4742 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.582808 4742 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.582876 4742 flags.go:64] FLAG: --system-cgroups="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.582939 4742 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.583024 4742 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.583091 4742 flags.go:64] FLAG: --tls-cert-file="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.583155 4742 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.583228 4742 flags.go:64] FLAG: --tls-min-version="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.583359 4742 flags.go:64] FLAG: --tls-private-key-file="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.583428 4742 flags.go:64] FLAG: --topology-manager-policy="none" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.583492 4742 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.583573 4742 flags.go:64] FLAG: --topology-manager-scope="container" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.583672 4742 flags.go:64] FLAG: --v="2" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.583758 4742 flags.go:64] FLAG: --version="false" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.583825 4742 flags.go:64] FLAG: --vmodule="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.583898 4742 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.583969 4742 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.584324 4742 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.584434 4742 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.584554 4742 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.584638 4742 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.584713 4742 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.584853 4742 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.584994 4742 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585070 4742 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585135 4742 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585225 4742 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585390 4742 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585419 4742 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585424 4742 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585429 4742 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585433 4742 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585438 4742 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585442 4742 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585452 4742 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585459 4742 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585464 4742 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585469 4742 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585477 4742 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585482 4742 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585486 4742 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585492 4742 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585496 4742 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585501 4742 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585505 4742 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585509 4742 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585513 4742 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585517 4742 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585521 4742 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585525 4742 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585528 4742 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585532 4742 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585535 4742 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585539 4742 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585543 4742 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585553 4742 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585558 4742 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585562 4742 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585567 4742 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585572 4742 feature_gate.go:330] unrecognized feature gate: Example Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585577 4742 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585582 4742 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585587 4742 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585591 4742 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585595 4742 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585599 4742 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585604 4742 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585610 4742 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585614 4742 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585618 4742 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585622 4742 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585626 4742 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585630 4742 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585634 4742 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585638 4742 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585642 4742 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585647 4742 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585652 4742 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585658 4742 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585662 4742 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585666 4742 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585670 4742 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585675 4742 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585678 4742 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585682 4742 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585687 4742 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585690 4742 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.585694 4742 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.585718 4742 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.596986 4742 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.597035 4742 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597169 4742 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597183 4742 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597192 4742 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597202 4742 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597211 4742 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597220 4742 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597230 4742 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597238 4742 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597247 4742 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597255 4742 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597263 4742 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597295 4742 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597303 4742 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597312 4742 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597319 4742 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597327 4742 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597335 4742 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597343 4742 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597351 4742 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597360 4742 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597369 4742 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597376 4742 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597385 4742 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597393 4742 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597401 4742 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597408 4742 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597416 4742 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597424 4742 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597431 4742 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597441 4742 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597450 4742 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597457 4742 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597465 4742 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597472 4742 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597480 4742 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597489 4742 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597496 4742 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597504 4742 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597512 4742 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597520 4742 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597529 4742 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597537 4742 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597546 4742 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597557 4742 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597583 4742 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597592 4742 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597600 4742 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597608 4742 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597616 4742 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597623 4742 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597632 4742 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597639 4742 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597647 4742 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597655 4742 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597665 4742 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597676 4742 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597686 4742 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597697 4742 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597707 4742 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597720 4742 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597731 4742 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597740 4742 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597753 4742 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597764 4742 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597773 4742 feature_gate.go:330] unrecognized feature gate: Example Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597782 4742 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597793 4742 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597804 4742 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597812 4742 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597820 4742 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.597828 4742 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.597843 4742 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598077 4742 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598089 4742 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598100 4742 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598109 4742 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598118 4742 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598126 4742 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598135 4742 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598143 4742 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598152 4742 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598161 4742 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598170 4742 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598180 4742 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598190 4742 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598199 4742 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598209 4742 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598218 4742 feature_gate.go:330] unrecognized feature gate: Example Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598228 4742 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598238 4742 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598247 4742 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598257 4742 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598266 4742 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598303 4742 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598311 4742 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598319 4742 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598327 4742 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598334 4742 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598342 4742 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598350 4742 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598357 4742 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598366 4742 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598377 4742 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598386 4742 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598394 4742 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598402 4742 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598410 4742 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598418 4742 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598426 4742 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598433 4742 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598441 4742 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598448 4742 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598459 4742 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598468 4742 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598476 4742 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598485 4742 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598493 4742 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598501 4742 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598509 4742 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598517 4742 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598524 4742 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598532 4742 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598540 4742 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598548 4742 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598604 4742 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598612 4742 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598622 4742 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598632 4742 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598641 4742 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598650 4742 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598659 4742 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598667 4742 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598710 4742 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598719 4742 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598727 4742 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598736 4742 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598744 4742 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598753 4742 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598762 4742 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598770 4742 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598777 4742 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598785 4742 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.598792 4742 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.598805 4742 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.599024 4742 server.go:940] "Client rotation is on, will bootstrap in background" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.605005 4742 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.605131 4742 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.609212 4742 server.go:997] "Starting client certificate rotation" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.609263 4742 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.609543 4742 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-12 00:23:12.785106319 +0000 UTC Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.609647 4742 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 2397h12m38.175463617s for next certificate rotation Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.645548 4742 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.648772 4742 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.673112 4742 log.go:25] "Validated CRI v1 runtime API" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.715217 4742 log.go:25] "Validated CRI v1 image API" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.718319 4742 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.729968 4742 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-04-03-06-10-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.730026 4742 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.760177 4742 manager.go:217] Machine: {Timestamp:2025-10-04 03:10:34.757446988 +0000 UTC m=+0.689784254 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:91a5b4ce-a0c3-485b-af77-0ed26accb46e BootID:94af7d13-529b-4a4d-a127-c697b50cfaf5 Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:45:87:bc Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:45:87:bc Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:5b:2d:ed Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:c1:ff:a9 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:ba:f5:c1 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:39:2c:e9 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:7e:dd:7e:b9:a1:72 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:3a:91:b8:4b:ea:d7 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.760645 4742 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.760848 4742 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.762608 4742 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.763038 4742 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.763105 4742 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.764334 4742 topology_manager.go:138] "Creating topology manager with none policy" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.764367 4742 container_manager_linux.go:303] "Creating device plugin manager" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.765025 4742 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.765065 4742 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.765337 4742 state_mem.go:36] "Initialized new in-memory state store" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.765476 4742 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.769921 4742 kubelet.go:418] "Attempting to sync node with API server" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.769968 4742 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.770005 4742 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.770030 4742 kubelet.go:324] "Adding apiserver pod source" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.770114 4742 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.775141 4742 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.776441 4742 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.777821 4742 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.189:6443: connect: connection refused Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.777909 4742 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 04 03:10:34 crc kubenswrapper[4742]: E1004 03:10:34.777940 4742 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.189:6443: connect: connection refused" logger="UnhandledError" Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.777827 4742 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.189:6443: connect: connection refused Oct 04 03:10:34 crc kubenswrapper[4742]: E1004 03:10:34.777993 4742 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.189:6443: connect: connection refused" logger="UnhandledError" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.779297 4742 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.779330 4742 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.779339 4742 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.779348 4742 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.779363 4742 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.779371 4742 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.779382 4742 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.779395 4742 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.779431 4742 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.779441 4742 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.779464 4742 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.779474 4742 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.781581 4742 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.782170 4742 server.go:1280] "Started kubelet" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.783998 4742 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.189:6443: connect: connection refused Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.784203 4742 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 04 03:10:34 crc systemd[1]: Started Kubernetes Kubelet. Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.784201 4742 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.784691 4742 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.785627 4742 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.785700 4742 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.785717 4742 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 21:19:53.144984906 +0000 UTC Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.785753 4742 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1410h9m18.35923482s for next certificate rotation Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.785947 4742 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.785968 4742 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 04 03:10:34 crc kubenswrapper[4742]: E1004 03:10:34.785961 4742 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.790552 4742 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.790907 4742 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.189:6443: connect: connection refused Oct 04 03:10:34 crc kubenswrapper[4742]: E1004 03:10:34.790974 4742 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.189:6443: connect: connection refused" logger="UnhandledError" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.791482 4742 factory.go:55] Registering systemd factory Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.791510 4742 factory.go:221] Registration of the systemd container factory successfully Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.792035 4742 factory.go:153] Registering CRI-O factory Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.792135 4742 factory.go:221] Registration of the crio container factory successfully Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.792548 4742 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.792798 4742 factory.go:103] Registering Raw factory Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.792837 4742 manager.go:1196] Started watching for new ooms in manager Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.794110 4742 manager.go:319] Starting recovery of all containers Oct 04 03:10:34 crc kubenswrapper[4742]: E1004 03:10:34.794405 4742 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.189:6443: connect: connection refused" interval="200ms" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.795255 4742 server.go:460] "Adding debug handlers to kubelet server" Oct 04 03:10:34 crc kubenswrapper[4742]: E1004 03:10:34.795908 4742 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.189:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186b2b11c7a38ae6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-04 03:10:34.782132966 +0000 UTC m=+0.714470152,LastTimestamp:2025-10-04 03:10:34.782132966 +0000 UTC m=+0.714470152,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.804661 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.804719 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.804733 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.804746 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.804758 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.804771 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.804784 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.804796 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.804814 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.804839 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.804856 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.804869 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.804883 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.804899 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.804922 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.804934 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.804947 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.804958 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.804971 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.804985 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.804999 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805012 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805025 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805037 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805053 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805066 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805082 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805095 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805109 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805125 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805137 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805151 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805164 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805177 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805189 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805201 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805214 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805227 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805239 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805249 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805263 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805298 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805310 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805325 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805340 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805351 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805368 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805381 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805393 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805406 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805421 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805434 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805453 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805467 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805482 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805496 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805511 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805524 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805537 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805550 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805562 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805576 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805588 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805601 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805627 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805642 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805655 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805667 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805679 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805692 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805707 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805720 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805764 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805778 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805789 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805803 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805814 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805826 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805838 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805853 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805866 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805877 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805889 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805901 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805912 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805922 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805934 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805945 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805959 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805970 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805982 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.805994 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.806005 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.806018 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.806030 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.806041 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.806052 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.806066 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.806079 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.806092 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.806103 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.806116 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.806129 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.806141 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.806165 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.806182 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.806195 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.806208 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.806221 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.806234 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.806251 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.806298 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.806315 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.806329 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.806343 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.806359 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.806370 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.806381 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.806394 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.806406 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.806419 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.806431 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.806442 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808059 4742 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808121 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808137 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808154 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808171 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808186 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808200 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808217 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808230 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808243 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808256 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808284 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808301 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808318 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808332 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808346 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808361 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808375 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808387 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808405 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808419 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808433 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808449 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808463 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808480 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808495 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808509 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808528 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808542 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808558 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808573 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808601 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808615 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808630 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808643 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808658 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808672 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808686 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808699 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808713 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808725 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808738 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808753 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808768 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808782 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808797 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808810 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808823 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808839 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808852 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808865 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808881 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808897 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808911 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808926 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808942 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808959 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808975 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.808989 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.809003 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.809018 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.809033 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.809046 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.809060 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.809075 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.809091 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.809104 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.809119 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.809133 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.809147 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.809163 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.809177 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.809193 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.809208 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.809227 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.809241 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.809255 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.809285 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.809302 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.809316 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.809333 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.809349 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.809364 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.809380 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.809395 4742 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.809407 4742 reconstruct.go:97] "Volume reconstruction finished" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.809416 4742 reconciler.go:26] "Reconciler: start to sync state" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.827027 4742 manager.go:324] Recovery completed Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.852377 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.852390 4742 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.855008 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.855063 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.855078 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.856351 4742 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.856377 4742 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.856400 4742 state_mem.go:36] "Initialized new in-memory state store" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.857045 4742 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.857108 4742 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.857148 4742 kubelet.go:2335] "Starting kubelet main sync loop" Oct 04 03:10:34 crc kubenswrapper[4742]: E1004 03:10:34.857243 4742 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 04 03:10:34 crc kubenswrapper[4742]: W1004 03:10:34.858143 4742 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.189:6443: connect: connection refused Oct 04 03:10:34 crc kubenswrapper[4742]: E1004 03:10:34.858242 4742 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.189:6443: connect: connection refused" logger="UnhandledError" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.882786 4742 policy_none.go:49] "None policy: Start" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.884010 4742 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.884074 4742 state_mem.go:35] "Initializing new in-memory state store" Oct 04 03:10:34 crc kubenswrapper[4742]: E1004 03:10:34.890703 4742 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.934830 4742 manager.go:334] "Starting Device Plugin manager" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.934917 4742 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.934932 4742 server.go:79] "Starting device plugin registration server" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.935489 4742 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.935604 4742 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.935858 4742 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.936002 4742 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.936013 4742 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 04 03:10:34 crc kubenswrapper[4742]: E1004 03:10:34.945714 4742 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.958172 4742 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.958260 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.959506 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.959539 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.959549 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.959678 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.960524 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.960547 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.960556 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.960867 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.960901 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.960859 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.960961 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.960968 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.961793 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.961845 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.961857 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.962088 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.962204 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.962246 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.962378 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.962420 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.962438 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.962760 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.962779 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.962789 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.963163 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.963187 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.963196 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.963303 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.963340 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.963351 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.963369 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.963541 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.963585 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.963993 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.964020 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.964031 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.964301 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.964337 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.964583 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.964610 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.964627 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.965201 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.965230 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:34 crc kubenswrapper[4742]: I1004 03:10:34.965241 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:34 crc kubenswrapper[4742]: E1004 03:10:34.995884 4742 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.189:6443: connect: connection refused" interval="400ms" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.013198 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.013250 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.013499 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.013527 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.013560 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.013600 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.013693 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.013742 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.013775 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.013880 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.013950 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.013985 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.014017 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.014046 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.014075 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.036076 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.039703 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.039771 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.039789 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.039823 4742 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 03:10:35 crc kubenswrapper[4742]: E1004 03:10:35.040649 4742 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.189:6443: connect: connection refused" node="crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.115973 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.116020 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.116039 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.116059 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.116076 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.116092 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.116105 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.116121 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.116138 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.116153 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.116168 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.116187 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.116184 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.116246 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.116204 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.116258 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.116308 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.116290 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.116290 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.116379 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.116332 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.116350 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.116357 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.116373 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.116381 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.116349 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.116392 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.116318 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.116588 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.116703 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.241054 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.242418 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.242462 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.242474 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.242533 4742 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 03:10:35 crc kubenswrapper[4742]: E1004 03:10:35.243354 4742 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.189:6443: connect: connection refused" node="crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.301945 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.317885 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.335637 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.350545 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.356223 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:10:35 crc kubenswrapper[4742]: W1004 03:10:35.365086 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-77bf3d023e3b88e085d73689c1c669bb2f84390f9f191fd34662956a0dfa2af1 WatchSource:0}: Error finding container 77bf3d023e3b88e085d73689c1c669bb2f84390f9f191fd34662956a0dfa2af1: Status 404 returned error can't find the container with id 77bf3d023e3b88e085d73689c1c669bb2f84390f9f191fd34662956a0dfa2af1 Oct 04 03:10:35 crc kubenswrapper[4742]: W1004 03:10:35.367337 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-53788f502985c1372d35f82e45235db314612a37374424adbbc45d94fe38100a WatchSource:0}: Error finding container 53788f502985c1372d35f82e45235db314612a37374424adbbc45d94fe38100a: Status 404 returned error can't find the container with id 53788f502985c1372d35f82e45235db314612a37374424adbbc45d94fe38100a Oct 04 03:10:35 crc kubenswrapper[4742]: W1004 03:10:35.379164 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-035f3d23ad1cd57904db7808bcef6b009d8fb3de0aeda0506388bd2112b37d1f WatchSource:0}: Error finding container 035f3d23ad1cd57904db7808bcef6b009d8fb3de0aeda0506388bd2112b37d1f: Status 404 returned error can't find the container with id 035f3d23ad1cd57904db7808bcef6b009d8fb3de0aeda0506388bd2112b37d1f Oct 04 03:10:35 crc kubenswrapper[4742]: W1004 03:10:35.381354 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-81d25be3a0361f99faf4d819ea3974caa7ff2fce0cf5e6e50a2e74c15ba95aee WatchSource:0}: Error finding container 81d25be3a0361f99faf4d819ea3974caa7ff2fce0cf5e6e50a2e74c15ba95aee: Status 404 returned error can't find the container with id 81d25be3a0361f99faf4d819ea3974caa7ff2fce0cf5e6e50a2e74c15ba95aee Oct 04 03:10:35 crc kubenswrapper[4742]: W1004 03:10:35.388791 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-0035f77ac58f77f2eafce7cf3265a769029ca2547edbac210ac5b2db888f23da WatchSource:0}: Error finding container 0035f77ac58f77f2eafce7cf3265a769029ca2547edbac210ac5b2db888f23da: Status 404 returned error can't find the container with id 0035f77ac58f77f2eafce7cf3265a769029ca2547edbac210ac5b2db888f23da Oct 04 03:10:35 crc kubenswrapper[4742]: E1004 03:10:35.397174 4742 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.189:6443: connect: connection refused" interval="800ms" Oct 04 03:10:35 crc kubenswrapper[4742]: W1004 03:10:35.607438 4742 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.189:6443: connect: connection refused Oct 04 03:10:35 crc kubenswrapper[4742]: E1004 03:10:35.607578 4742 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.189:6443: connect: connection refused" logger="UnhandledError" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.643966 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.645493 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.645583 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.645599 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.645643 4742 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 03:10:35 crc kubenswrapper[4742]: E1004 03:10:35.646373 4742 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.189:6443: connect: connection refused" node="crc" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.785683 4742 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.189:6443: connect: connection refused Oct 04 03:10:35 crc kubenswrapper[4742]: W1004 03:10:35.803913 4742 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.189:6443: connect: connection refused Oct 04 03:10:35 crc kubenswrapper[4742]: E1004 03:10:35.804012 4742 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.189:6443: connect: connection refused" logger="UnhandledError" Oct 04 03:10:35 crc kubenswrapper[4742]: W1004 03:10:35.840043 4742 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.189:6443: connect: connection refused Oct 04 03:10:35 crc kubenswrapper[4742]: E1004 03:10:35.840130 4742 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.189:6443: connect: connection refused" logger="UnhandledError" Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.861935 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"81d25be3a0361f99faf4d819ea3974caa7ff2fce0cf5e6e50a2e74c15ba95aee"} Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.863195 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"035f3d23ad1cd57904db7808bcef6b009d8fb3de0aeda0506388bd2112b37d1f"} Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.864256 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"53788f502985c1372d35f82e45235db314612a37374424adbbc45d94fe38100a"} Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.865210 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"77bf3d023e3b88e085d73689c1c669bb2f84390f9f191fd34662956a0dfa2af1"} Oct 04 03:10:35 crc kubenswrapper[4742]: I1004 03:10:35.866513 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0035f77ac58f77f2eafce7cf3265a769029ca2547edbac210ac5b2db888f23da"} Oct 04 03:10:36 crc kubenswrapper[4742]: E1004 03:10:36.198909 4742 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.189:6443: connect: connection refused" interval="1.6s" Oct 04 03:10:36 crc kubenswrapper[4742]: W1004 03:10:36.243666 4742 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.189:6443: connect: connection refused Oct 04 03:10:36 crc kubenswrapper[4742]: E1004 03:10:36.243768 4742 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.189:6443: connect: connection refused" logger="UnhandledError" Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.447239 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.449685 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.449759 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.449777 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.449818 4742 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 03:10:36 crc kubenswrapper[4742]: E1004 03:10:36.450593 4742 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.189:6443: connect: connection refused" node="crc" Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.785831 4742 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.189:6443: connect: connection refused Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.874815 4742 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="35257b186eeb8544b8912f2a2c995eccd0549a4be771fa68a17b6489fc5dc7ef" exitCode=0 Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.874900 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"35257b186eeb8544b8912f2a2c995eccd0549a4be771fa68a17b6489fc5dc7ef"} Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.875030 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.876411 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.876468 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.876482 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.877713 4742 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="ce3ee868ef16effadb77812f1673237d65d10493bd29dfcef5c7fe53da640d4e" exitCode=0 Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.877764 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"ce3ee868ef16effadb77812f1673237d65d10493bd29dfcef5c7fe53da640d4e"} Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.877796 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.878830 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.878898 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.878915 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.880270 4742 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1" exitCode=0 Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.880364 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1"} Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.880396 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.881948 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.881988 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.882002 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.883795 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.884852 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.884906 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.884919 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.886272 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51"} Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.886338 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0"} Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.886352 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad"} Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.886453 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.888083 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.888121 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.888131 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.889867 4742 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="48906a72bb35674544fa9d074de2f582da4cbad41e0472f6bea749cd6a7d860c" exitCode=0 Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.889924 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"48906a72bb35674544fa9d074de2f582da4cbad41e0472f6bea749cd6a7d860c"} Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.889970 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.891352 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.891398 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:36 crc kubenswrapper[4742]: I1004 03:10:36.891410 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:37 crc kubenswrapper[4742]: W1004 03:10:37.314629 4742 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.189:6443: connect: connection refused Oct 04 03:10:37 crc kubenswrapper[4742]: E1004 03:10:37.314759 4742 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.189:6443: connect: connection refused" logger="UnhandledError" Oct 04 03:10:37 crc kubenswrapper[4742]: W1004 03:10:37.710252 4742 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.189:6443: connect: connection refused Oct 04 03:10:37 crc kubenswrapper[4742]: E1004 03:10:37.710382 4742 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.189:6443: connect: connection refused" logger="UnhandledError" Oct 04 03:10:37 crc kubenswrapper[4742]: E1004 03:10:37.715911 4742 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.189:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186b2b11c7a38ae6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-04 03:10:34.782132966 +0000 UTC m=+0.714470152,LastTimestamp:2025-10-04 03:10:34.782132966 +0000 UTC m=+0.714470152,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 04 03:10:37 crc kubenswrapper[4742]: I1004 03:10:37.784904 4742 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.189:6443: connect: connection refused Oct 04 03:10:37 crc kubenswrapper[4742]: E1004 03:10:37.800216 4742 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.189:6443: connect: connection refused" interval="3.2s" Oct 04 03:10:37 crc kubenswrapper[4742]: I1004 03:10:37.894530 4742 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="c9bcffac1019b9253a814cd8b0951819a40309057f15c54bf6ff4ee0fc491e61" exitCode=0 Oct 04 03:10:37 crc kubenswrapper[4742]: I1004 03:10:37.894653 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"c9bcffac1019b9253a814cd8b0951819a40309057f15c54bf6ff4ee0fc491e61"} Oct 04 03:10:37 crc kubenswrapper[4742]: I1004 03:10:37.894683 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:37 crc kubenswrapper[4742]: I1004 03:10:37.895892 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:37 crc kubenswrapper[4742]: I1004 03:10:37.895931 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:37 crc kubenswrapper[4742]: I1004 03:10:37.895946 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:37 crc kubenswrapper[4742]: I1004 03:10:37.896595 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"c7871a801b96fa8ef62e65a229e503e2914a008f35163b85cbffc99ffc543c37"} Oct 04 03:10:37 crc kubenswrapper[4742]: I1004 03:10:37.896669 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:37 crc kubenswrapper[4742]: I1004 03:10:37.897404 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:37 crc kubenswrapper[4742]: I1004 03:10:37.897439 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:37 crc kubenswrapper[4742]: I1004 03:10:37.897454 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:37 crc kubenswrapper[4742]: I1004 03:10:37.899441 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"f2ede8f31552e98a3087393a27a4bab2c56f116829fef768b9c1dda8d6630142"} Oct 04 03:10:37 crc kubenswrapper[4742]: I1004 03:10:37.899464 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:37 crc kubenswrapper[4742]: I1004 03:10:37.899474 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ac618416ef312d2c051d88c0348fea230d73d3c8d6f4b1830b4b584ceae645c8"} Oct 04 03:10:37 crc kubenswrapper[4742]: I1004 03:10:37.899488 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"bf178a15abcb94eda5033b74ef132d9e204b33f4112152edf172e61999fc2c9d"} Oct 04 03:10:37 crc kubenswrapper[4742]: I1004 03:10:37.900223 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:37 crc kubenswrapper[4742]: I1004 03:10:37.900268 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:37 crc kubenswrapper[4742]: I1004 03:10:37.900301 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:37 crc kubenswrapper[4742]: I1004 03:10:37.902574 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890"} Oct 04 03:10:37 crc kubenswrapper[4742]: I1004 03:10:37.902611 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a"} Oct 04 03:10:37 crc kubenswrapper[4742]: I1004 03:10:37.902629 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef"} Oct 04 03:10:37 crc kubenswrapper[4742]: I1004 03:10:37.902647 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11"} Oct 04 03:10:37 crc kubenswrapper[4742]: I1004 03:10:37.904954 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196"} Oct 04 03:10:37 crc kubenswrapper[4742]: I1004 03:10:37.905033 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:37 crc kubenswrapper[4742]: I1004 03:10:37.905813 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:37 crc kubenswrapper[4742]: I1004 03:10:37.905836 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:37 crc kubenswrapper[4742]: I1004 03:10:37.905845 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:38 crc kubenswrapper[4742]: I1004 03:10:38.051473 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:38 crc kubenswrapper[4742]: I1004 03:10:38.053207 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:38 crc kubenswrapper[4742]: I1004 03:10:38.053261 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:38 crc kubenswrapper[4742]: I1004 03:10:38.053372 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:38 crc kubenswrapper[4742]: I1004 03:10:38.053432 4742 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 03:10:38 crc kubenswrapper[4742]: E1004 03:10:38.054173 4742 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.189:6443: connect: connection refused" node="crc" Oct 04 03:10:38 crc kubenswrapper[4742]: W1004 03:10:38.160073 4742 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.189:6443: connect: connection refused Oct 04 03:10:38 crc kubenswrapper[4742]: E1004 03:10:38.160209 4742 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.189:6443: connect: connection refused" logger="UnhandledError" Oct 04 03:10:38 crc kubenswrapper[4742]: I1004 03:10:38.914102 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610"} Oct 04 03:10:38 crc kubenswrapper[4742]: I1004 03:10:38.914378 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:38 crc kubenswrapper[4742]: I1004 03:10:38.915994 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:38 crc kubenswrapper[4742]: I1004 03:10:38.916058 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:38 crc kubenswrapper[4742]: I1004 03:10:38.916079 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:38 crc kubenswrapper[4742]: I1004 03:10:38.916707 4742 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="0fd66b11b8abc5ab89eb8355cba078dc46e2be3ee0f59afc37cdcee8e80115b1" exitCode=0 Oct 04 03:10:38 crc kubenswrapper[4742]: I1004 03:10:38.916761 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"0fd66b11b8abc5ab89eb8355cba078dc46e2be3ee0f59afc37cdcee8e80115b1"} Oct 04 03:10:38 crc kubenswrapper[4742]: I1004 03:10:38.916829 4742 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 03:10:38 crc kubenswrapper[4742]: I1004 03:10:38.916842 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:38 crc kubenswrapper[4742]: I1004 03:10:38.916898 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:38 crc kubenswrapper[4742]: I1004 03:10:38.916958 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:38 crc kubenswrapper[4742]: I1004 03:10:38.917585 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:38 crc kubenswrapper[4742]: I1004 03:10:38.918324 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:38 crc kubenswrapper[4742]: I1004 03:10:38.918507 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:38 crc kubenswrapper[4742]: I1004 03:10:38.918603 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:38 crc kubenswrapper[4742]: I1004 03:10:38.918505 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:38 crc kubenswrapper[4742]: I1004 03:10:38.918749 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:38 crc kubenswrapper[4742]: I1004 03:10:38.918777 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:38 crc kubenswrapper[4742]: I1004 03:10:38.918411 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:38 crc kubenswrapper[4742]: I1004 03:10:38.918831 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:38 crc kubenswrapper[4742]: I1004 03:10:38.918845 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:38 crc kubenswrapper[4742]: I1004 03:10:38.919011 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:38 crc kubenswrapper[4742]: I1004 03:10:38.919040 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:38 crc kubenswrapper[4742]: I1004 03:10:38.919052 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:39 crc kubenswrapper[4742]: I1004 03:10:39.410382 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:10:39 crc kubenswrapper[4742]: I1004 03:10:39.435896 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:10:39 crc kubenswrapper[4742]: I1004 03:10:39.923947 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"01061b91a5279d579507bc180cc59378d3729dc8d327f90a37f3e89331d6a38d"} Oct 04 03:10:39 crc kubenswrapper[4742]: I1004 03:10:39.924011 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"65ba85e90248976f60214ebcda8e19348e9c373aa509ca521cb8e313acec84c1"} Oct 04 03:10:39 crc kubenswrapper[4742]: I1004 03:10:39.924027 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"830fe4584483d987ccd84235d0d565579604c385ffc648b6a61868a7369a19c9"} Oct 04 03:10:39 crc kubenswrapper[4742]: I1004 03:10:39.924038 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"467455dfc35412e2c98ab78be938378877791ef8e6acd843795dd5e600495934"} Oct 04 03:10:39 crc kubenswrapper[4742]: I1004 03:10:39.924053 4742 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 03:10:39 crc kubenswrapper[4742]: I1004 03:10:39.924101 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:39 crc kubenswrapper[4742]: I1004 03:10:39.924115 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:39 crc kubenswrapper[4742]: I1004 03:10:39.925296 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:39 crc kubenswrapper[4742]: I1004 03:10:39.925329 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:39 crc kubenswrapper[4742]: I1004 03:10:39.925338 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:39 crc kubenswrapper[4742]: I1004 03:10:39.925393 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:39 crc kubenswrapper[4742]: I1004 03:10:39.925360 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:39 crc kubenswrapper[4742]: I1004 03:10:39.925433 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:40 crc kubenswrapper[4742]: I1004 03:10:40.935245 4742 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 03:10:40 crc kubenswrapper[4742]: I1004 03:10:40.935306 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"0f53f64a2c0a1b813ae42baac78f5e639dfc576212d66b48480b0d3563120e12"} Oct 04 03:10:40 crc kubenswrapper[4742]: I1004 03:10:40.935397 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:40 crc kubenswrapper[4742]: I1004 03:10:40.935527 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:40 crc kubenswrapper[4742]: I1004 03:10:40.937001 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:40 crc kubenswrapper[4742]: I1004 03:10:40.937036 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:40 crc kubenswrapper[4742]: I1004 03:10:40.937049 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:40 crc kubenswrapper[4742]: I1004 03:10:40.938027 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:40 crc kubenswrapper[4742]: I1004 03:10:40.938098 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:40 crc kubenswrapper[4742]: I1004 03:10:40.938120 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:41 crc kubenswrapper[4742]: I1004 03:10:41.254955 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:41 crc kubenswrapper[4742]: I1004 03:10:41.257003 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:41 crc kubenswrapper[4742]: I1004 03:10:41.257072 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:41 crc kubenswrapper[4742]: I1004 03:10:41.257091 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:41 crc kubenswrapper[4742]: I1004 03:10:41.257129 4742 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 03:10:41 crc kubenswrapper[4742]: I1004 03:10:41.283900 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:10:41 crc kubenswrapper[4742]: I1004 03:10:41.284192 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:41 crc kubenswrapper[4742]: I1004 03:10:41.286219 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:41 crc kubenswrapper[4742]: I1004 03:10:41.286312 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:41 crc kubenswrapper[4742]: I1004 03:10:41.286331 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:41 crc kubenswrapper[4742]: I1004 03:10:41.939128 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:41 crc kubenswrapper[4742]: I1004 03:10:41.940713 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:41 crc kubenswrapper[4742]: I1004 03:10:41.940802 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:41 crc kubenswrapper[4742]: I1004 03:10:41.940826 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:42 crc kubenswrapper[4742]: I1004 03:10:42.304315 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:10:42 crc kubenswrapper[4742]: I1004 03:10:42.304510 4742 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 03:10:42 crc kubenswrapper[4742]: I1004 03:10:42.304554 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:42 crc kubenswrapper[4742]: I1004 03:10:42.305877 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:42 crc kubenswrapper[4742]: I1004 03:10:42.305912 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:42 crc kubenswrapper[4742]: I1004 03:10:42.305926 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:44 crc kubenswrapper[4742]: I1004 03:10:44.804399 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 03:10:44 crc kubenswrapper[4742]: I1004 03:10:44.804687 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:44 crc kubenswrapper[4742]: I1004 03:10:44.806554 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:44 crc kubenswrapper[4742]: I1004 03:10:44.806628 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:44 crc kubenswrapper[4742]: I1004 03:10:44.806644 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:44 crc kubenswrapper[4742]: E1004 03:10:44.946511 4742 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 04 03:10:45 crc kubenswrapper[4742]: I1004 03:10:45.078929 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:10:45 crc kubenswrapper[4742]: I1004 03:10:45.079139 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:45 crc kubenswrapper[4742]: I1004 03:10:45.080617 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:45 crc kubenswrapper[4742]: I1004 03:10:45.080709 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:45 crc kubenswrapper[4742]: I1004 03:10:45.080729 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:45 crc kubenswrapper[4742]: I1004 03:10:45.112175 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:10:45 crc kubenswrapper[4742]: I1004 03:10:45.112452 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:45 crc kubenswrapper[4742]: I1004 03:10:45.114121 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:45 crc kubenswrapper[4742]: I1004 03:10:45.114162 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:45 crc kubenswrapper[4742]: I1004 03:10:45.114171 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:45 crc kubenswrapper[4742]: I1004 03:10:45.606180 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 04 03:10:45 crc kubenswrapper[4742]: I1004 03:10:45.606370 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:45 crc kubenswrapper[4742]: I1004 03:10:45.608339 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:45 crc kubenswrapper[4742]: I1004 03:10:45.608410 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:45 crc kubenswrapper[4742]: I1004 03:10:45.608431 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:46 crc kubenswrapper[4742]: I1004 03:10:46.817340 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:10:46 crc kubenswrapper[4742]: I1004 03:10:46.817524 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:46 crc kubenswrapper[4742]: I1004 03:10:46.818872 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:46 crc kubenswrapper[4742]: I1004 03:10:46.818923 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:46 crc kubenswrapper[4742]: I1004 03:10:46.818934 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:46 crc kubenswrapper[4742]: I1004 03:10:46.824250 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:10:46 crc kubenswrapper[4742]: I1004 03:10:46.954615 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:46 crc kubenswrapper[4742]: I1004 03:10:46.956714 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:46 crc kubenswrapper[4742]: I1004 03:10:46.956853 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:46 crc kubenswrapper[4742]: I1004 03:10:46.956961 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:46 crc kubenswrapper[4742]: I1004 03:10:46.962692 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:10:47 crc kubenswrapper[4742]: I1004 03:10:47.961311 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:47 crc kubenswrapper[4742]: I1004 03:10:47.965605 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:47 crc kubenswrapper[4742]: I1004 03:10:47.965650 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:47 crc kubenswrapper[4742]: I1004 03:10:47.965660 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:48 crc kubenswrapper[4742]: I1004 03:10:48.113119 4742 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 04 03:10:48 crc kubenswrapper[4742]: I1004 03:10:48.113200 4742 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 03:10:48 crc kubenswrapper[4742]: I1004 03:10:48.786826 4742 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 04 03:10:48 crc kubenswrapper[4742]: I1004 03:10:48.990411 4742 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 04 03:10:48 crc kubenswrapper[4742]: I1004 03:10:48.990891 4742 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 04 03:10:48 crc kubenswrapper[4742]: I1004 03:10:48.995617 4742 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 04 03:10:48 crc kubenswrapper[4742]: I1004 03:10:48.995674 4742 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 04 03:10:49 crc kubenswrapper[4742]: I1004 03:10:49.420086 4742 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 04 03:10:49 crc kubenswrapper[4742]: [+]log ok Oct 04 03:10:49 crc kubenswrapper[4742]: [+]etcd ok Oct 04 03:10:49 crc kubenswrapper[4742]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 04 03:10:49 crc kubenswrapper[4742]: [+]poststarthook/openshift.io-api-request-count-filter ok Oct 04 03:10:49 crc kubenswrapper[4742]: [+]poststarthook/openshift.io-startkubeinformers ok Oct 04 03:10:49 crc kubenswrapper[4742]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Oct 04 03:10:49 crc kubenswrapper[4742]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Oct 04 03:10:49 crc kubenswrapper[4742]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 04 03:10:49 crc kubenswrapper[4742]: [+]poststarthook/generic-apiserver-start-informers ok Oct 04 03:10:49 crc kubenswrapper[4742]: [+]poststarthook/priority-and-fairness-config-consumer ok Oct 04 03:10:49 crc kubenswrapper[4742]: [+]poststarthook/priority-and-fairness-filter ok Oct 04 03:10:49 crc kubenswrapper[4742]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 04 03:10:49 crc kubenswrapper[4742]: [+]poststarthook/start-apiextensions-informers ok Oct 04 03:10:49 crc kubenswrapper[4742]: [+]poststarthook/start-apiextensions-controllers ok Oct 04 03:10:49 crc kubenswrapper[4742]: [+]poststarthook/crd-informer-synced ok Oct 04 03:10:49 crc kubenswrapper[4742]: [+]poststarthook/start-system-namespaces-controller ok Oct 04 03:10:49 crc kubenswrapper[4742]: [+]poststarthook/start-cluster-authentication-info-controller ok Oct 04 03:10:49 crc kubenswrapper[4742]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Oct 04 03:10:49 crc kubenswrapper[4742]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Oct 04 03:10:49 crc kubenswrapper[4742]: [+]poststarthook/start-legacy-token-tracking-controller ok Oct 04 03:10:49 crc kubenswrapper[4742]: [+]poststarthook/start-service-ip-repair-controllers ok Oct 04 03:10:49 crc kubenswrapper[4742]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Oct 04 03:10:49 crc kubenswrapper[4742]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Oct 04 03:10:49 crc kubenswrapper[4742]: [+]poststarthook/priority-and-fairness-config-producer ok Oct 04 03:10:49 crc kubenswrapper[4742]: [+]poststarthook/bootstrap-controller ok Oct 04 03:10:49 crc kubenswrapper[4742]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Oct 04 03:10:49 crc kubenswrapper[4742]: [+]poststarthook/start-kube-aggregator-informers ok Oct 04 03:10:49 crc kubenswrapper[4742]: [+]poststarthook/apiservice-status-local-available-controller ok Oct 04 03:10:49 crc kubenswrapper[4742]: [+]poststarthook/apiservice-status-remote-available-controller ok Oct 04 03:10:49 crc kubenswrapper[4742]: [+]poststarthook/apiservice-registration-controller ok Oct 04 03:10:49 crc kubenswrapper[4742]: [+]poststarthook/apiservice-wait-for-first-sync ok Oct 04 03:10:49 crc kubenswrapper[4742]: [+]poststarthook/apiservice-discovery-controller ok Oct 04 03:10:49 crc kubenswrapper[4742]: [+]poststarthook/kube-apiserver-autoregistration ok Oct 04 03:10:49 crc kubenswrapper[4742]: [+]autoregister-completion ok Oct 04 03:10:49 crc kubenswrapper[4742]: [+]poststarthook/apiservice-openapi-controller ok Oct 04 03:10:49 crc kubenswrapper[4742]: [+]poststarthook/apiservice-openapiv3-controller ok Oct 04 03:10:49 crc kubenswrapper[4742]: livez check failed Oct 04 03:10:49 crc kubenswrapper[4742]: I1004 03:10:49.420164 4742 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:10:50 crc kubenswrapper[4742]: I1004 03:10:50.033979 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 04 03:10:50 crc kubenswrapper[4742]: I1004 03:10:50.034211 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:50 crc kubenswrapper[4742]: I1004 03:10:50.035572 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:50 crc kubenswrapper[4742]: I1004 03:10:50.035637 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:50 crc kubenswrapper[4742]: I1004 03:10:50.035649 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:50 crc kubenswrapper[4742]: I1004 03:10:50.111434 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 04 03:10:50 crc kubenswrapper[4742]: I1004 03:10:50.623173 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 04 03:10:50 crc kubenswrapper[4742]: I1004 03:10:50.968410 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:50 crc kubenswrapper[4742]: I1004 03:10:50.969161 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:50 crc kubenswrapper[4742]: I1004 03:10:50.969200 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:50 crc kubenswrapper[4742]: I1004 03:10:50.969212 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:51 crc kubenswrapper[4742]: I1004 03:10:51.970258 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:10:51 crc kubenswrapper[4742]: I1004 03:10:51.971094 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:10:51 crc kubenswrapper[4742]: I1004 03:10:51.971129 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:10:51 crc kubenswrapper[4742]: I1004 03:10:51.971139 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:10:53 crc kubenswrapper[4742]: E1004 03:10:53.992962 4742 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 04 03:10:53 crc kubenswrapper[4742]: I1004 03:10:53.997475 4742 trace.go:236] Trace[532700783]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Oct-2025 03:10:41.865) (total time: 12131ms): Oct 04 03:10:53 crc kubenswrapper[4742]: Trace[532700783]: ---"Objects listed" error: 12131ms (03:10:53.997) Oct 04 03:10:53 crc kubenswrapper[4742]: Trace[532700783]: [12.131777202s] [12.131777202s] END Oct 04 03:10:53 crc kubenswrapper[4742]: I1004 03:10:53.997512 4742 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 04 03:10:53 crc kubenswrapper[4742]: I1004 03:10:53.999520 4742 trace.go:236] Trace[1070305456]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Oct-2025 03:10:39.426) (total time: 14572ms): Oct 04 03:10:53 crc kubenswrapper[4742]: Trace[1070305456]: ---"Objects listed" error: 14572ms (03:10:53.999) Oct 04 03:10:53 crc kubenswrapper[4742]: Trace[1070305456]: [14.572642865s] [14.572642865s] END Oct 04 03:10:53 crc kubenswrapper[4742]: I1004 03:10:53.999583 4742 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.000100 4742 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.001730 4742 trace.go:236] Trace[804756582]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Oct-2025 03:10:42.665) (total time: 11336ms): Oct 04 03:10:54 crc kubenswrapper[4742]: Trace[804756582]: ---"Objects listed" error: 11336ms (03:10:54.001) Oct 04 03:10:54 crc kubenswrapper[4742]: Trace[804756582]: [11.336042783s] [11.336042783s] END Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.001928 4742 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.003483 4742 trace.go:236] Trace[828578529]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Oct-2025 03:10:43.592) (total time: 10410ms): Oct 04 03:10:54 crc kubenswrapper[4742]: Trace[828578529]: ---"Objects listed" error: 10410ms (03:10:54.003) Oct 04 03:10:54 crc kubenswrapper[4742]: Trace[828578529]: [10.410435128s] [10.410435128s] END Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.003517 4742 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 04 03:10:54 crc kubenswrapper[4742]: E1004 03:10:54.005943 4742 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.414986 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.415537 4742 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.415643 4742 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.420104 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.779620 4742 apiserver.go:52] "Watching apiserver" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.788059 4742 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.788452 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.788849 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.789071 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.789217 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:10:54 crc kubenswrapper[4742]: E1004 03:10:54.789225 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:10:54 crc kubenswrapper[4742]: E1004 03:10:54.789297 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.789368 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.789383 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.789428 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:10:54 crc kubenswrapper[4742]: E1004 03:10:54.789892 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.791768 4742 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.794093 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.794331 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.794624 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.794650 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.794646 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.794779 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.794838 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.795027 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.795122 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.804464 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.804503 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.804538 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.804567 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.804927 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.805354 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.805430 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.806294 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.805076 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.805506 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.806322 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.805970 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.806138 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.806481 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.806543 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.806629 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.806658 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.806690 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.806715 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.806748 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.806778 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.806801 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.806826 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.806847 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.806843 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: E1004 03:10:54.806895 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:10:55.306847162 +0000 UTC m=+21.239184548 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.806956 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.807024 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.807082 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.807157 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.807216 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.807318 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.807372 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.807426 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.807484 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.807535 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.807595 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.807649 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.807701 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.807753 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.807785 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.807809 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.807870 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.807929 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.807982 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.808041 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.808092 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.808149 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.808207 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.808319 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.808394 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.808451 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.808487 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.808543 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.808522 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.808647 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.808710 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.808741 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.808760 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.808780 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.808803 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.808821 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.808839 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.808861 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.808849 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.808880 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.809087 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.809145 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.809189 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.809232 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.809300 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.809320 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.809354 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.809337 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.809451 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.809480 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.809649 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.809729 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.809783 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.809873 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.809912 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.809935 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.809988 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.810042 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.810059 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.810103 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.810160 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.810216 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.810265 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.810369 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.810422 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.810482 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.810537 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.810593 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.810661 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.810723 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.810776 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.810824 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.810874 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.810954 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.811025 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.811100 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.811162 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.811222 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.811362 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.811424 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.811476 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.811526 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.811631 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.811685 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.811737 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.811787 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.811839 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.811893 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.811954 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.812008 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.812057 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.812134 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.812196 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.812258 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.812389 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.812445 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.812498 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.812558 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.812677 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.812732 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.812786 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.812852 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.812917 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.812976 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.813030 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.813083 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.813141 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.813249 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.813376 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.813437 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.813497 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.813546 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.813600 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.813660 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.813711 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.813775 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.813868 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.813935 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.813995 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.814036 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.814073 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.814117 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.814163 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.814232 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.814328 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.814394 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.814456 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.814529 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.814598 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.814653 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.814706 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.814781 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.814843 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.814906 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.814973 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.815039 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.815101 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.815161 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.815217 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.815307 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.815366 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.815431 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.815499 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.815578 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.815648 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.815705 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.815768 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.815827 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.815888 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.815945 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.816011 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.816071 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.816133 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.816198 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.816263 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.816365 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.816423 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.816505 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.816571 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.816627 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.816677 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.816742 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.816801 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.817151 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.817311 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.817370 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.817430 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.817492 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.817554 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.817628 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.817705 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.817781 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.817880 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.817945 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.818001 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.818060 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.818122 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.818198 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.818257 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.818382 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.818446 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.818505 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.818568 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.818620 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.818674 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.818742 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.818845 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.818928 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.819023 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.819097 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.819180 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.819258 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.819367 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.819412 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.819464 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.819511 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.819557 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.819598 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.819643 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.819736 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.819854 4742 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.819889 4742 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.819923 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.819957 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.819988 4742 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.820018 4742 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.820046 4742 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.820078 4742 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.820113 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.820143 4742 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.820172 4742 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.820200 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.820229 4742 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.820263 4742 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.820329 4742 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.820361 4742 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.810968 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.811591 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.811972 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.812495 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.812584 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.824171 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.812663 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.812684 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.812731 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.812763 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.812803 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.813070 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.814324 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.814450 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.814696 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.814802 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.815444 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.815647 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.815635 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.816055 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.816226 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.817056 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.817091 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.817582 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.818066 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.818088 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.818554 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.818582 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.818905 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.819026 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.819457 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.819489 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.819578 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.819654 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.819664 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.819669 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.819703 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: E1004 03:10:54.820569 4742 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.820955 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.821004 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.821664 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.821753 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.821780 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.821937 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.822027 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.821009 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.822070 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.822102 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.822208 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.822221 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.822670 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.822749 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.822798 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.822830 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.822869 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.823143 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.823292 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.823305 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.823550 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.824019 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.824133 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.824301 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.824593 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.825174 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.825326 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.825433 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.825644 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.825647 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.825724 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.825786 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.826747 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.826038 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.826899 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.826952 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: E1004 03:10:54.827018 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:10:55.326987055 +0000 UTC m=+21.259324491 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.827118 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.827466 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.827561 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.827623 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.827625 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.827905 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.827926 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.827950 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.828563 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.828422 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.828841 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.828867 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.828965 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.829044 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.829057 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.829152 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.829153 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.829209 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.829231 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.829600 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.829683 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.830012 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.830144 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.830478 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.830493 4742 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.830694 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.831039 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.831181 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.831375 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.831493 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.831544 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.831568 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.831758 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.831903 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.831882 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.831939 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.832091 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: E1004 03:10:54.832183 4742 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:10:54 crc kubenswrapper[4742]: E1004 03:10:54.832593 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:10:55.33255553 +0000 UTC m=+21.264892926 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.832261 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.832314 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.832767 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.832872 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.833171 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.833582 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.833626 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.834971 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.835810 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.836485 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.836817 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.837151 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.837168 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.837474 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.837603 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.838065 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.838219 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.838400 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.838771 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.838850 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.838920 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.838945 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.839016 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.840037 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.844017 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.844196 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.848771 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.848954 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.849398 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.851548 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.852128 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.852801 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: E1004 03:10:54.853053 4742 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:10:54 crc kubenswrapper[4742]: E1004 03:10:54.853113 4742 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:10:54 crc kubenswrapper[4742]: E1004 03:10:54.853146 4742 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:10:54 crc kubenswrapper[4742]: E1004 03:10:54.853306 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 03:10:55.353232357 +0000 UTC m=+21.285569703 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:10:54 crc kubenswrapper[4742]: E1004 03:10:54.853680 4742 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:10:54 crc kubenswrapper[4742]: E1004 03:10:54.853830 4742 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:10:54 crc kubenswrapper[4742]: E1004 03:10:54.853977 4742 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.854119 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.854309 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.854378 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.854447 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.854453 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.854002 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.854627 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.854937 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:10:54 crc kubenswrapper[4742]: E1004 03:10:54.855082 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 03:10:55.354237653 +0000 UTC m=+21.286574859 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.855135 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.855243 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.855339 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.855587 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.855635 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.855759 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.855928 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.856082 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.856164 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.856199 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.856324 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.856587 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.856620 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.857116 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.857133 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.857390 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.857473 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.857526 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.857720 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.858177 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.858313 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.858589 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.858792 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.859408 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.859863 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.860887 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.861261 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.861374 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.861611 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.862939 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.867718 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.868003 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.868092 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.868337 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.870015 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.876092 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.877825 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.879801 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.879865 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.881555 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.884407 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.884613 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.886534 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.887479 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.888441 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.888853 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.889769 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.890694 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.892529 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.893400 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.894737 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.895795 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.897359 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.898164 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.899562 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.900380 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.901587 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.902117 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.902230 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.903085 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.904798 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.905890 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.906632 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.908093 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.908729 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.910217 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.910974 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.911921 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.913479 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.914223 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.914498 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.915662 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.916429 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.917134 4742 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.917305 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.919319 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.919975 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.921460 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.921658 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.921714 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.922156 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.922363 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.922397 4742 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.922428 4742 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.922449 4742 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.922470 4742 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.922490 4742 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.922506 4742 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.922525 4742 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.922543 4742 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.922560 4742 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.922577 4742 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.922595 4742 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.922621 4742 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.922641 4742 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.922661 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.922678 4742 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.922696 4742 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.922713 4742 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.922730 4742 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.922747 4742 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.922764 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.922782 4742 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.922799 4742 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.922816 4742 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.922836 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.922853 4742 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.922868 4742 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.922886 4742 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.922903 4742 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.922923 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.922941 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.922961 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.922979 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.922996 4742 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923014 4742 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923031 4742 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923048 4742 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923065 4742 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923082 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923101 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923118 4742 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923134 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923148 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923163 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923177 4742 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923192 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923208 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923223 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923241 4742 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923257 4742 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923302 4742 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923320 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923337 4742 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923353 4742 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923369 4742 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923387 4742 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923406 4742 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923421 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923438 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923456 4742 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923472 4742 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923487 4742 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923507 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923524 4742 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923718 4742 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923742 4742 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923779 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923792 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923804 4742 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923817 4742 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923839 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923866 4742 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923848 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.923882 4742 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924102 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924114 4742 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924125 4742 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924135 4742 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924144 4742 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924155 4742 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924163 4742 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924173 4742 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924183 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924192 4742 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924202 4742 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924213 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924223 4742 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924232 4742 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924242 4742 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924251 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924260 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924291 4742 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924300 4742 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924309 4742 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924318 4742 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924327 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924337 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924346 4742 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924355 4742 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924364 4742 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924376 4742 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924386 4742 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924395 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924405 4742 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924415 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924426 4742 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924435 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924444 4742 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924452 4742 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924462 4742 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924470 4742 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924479 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924488 4742 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924497 4742 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924506 4742 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924515 4742 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924523 4742 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924532 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924541 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924550 4742 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924558 4742 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924567 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924575 4742 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924583 4742 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924592 4742 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924601 4742 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924611 4742 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924622 4742 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924632 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924642 4742 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924650 4742 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924659 4742 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924666 4742 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924675 4742 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924683 4742 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924691 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924700 4742 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924710 4742 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924718 4742 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924726 4742 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924734 4742 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924743 4742 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924752 4742 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924761 4742 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924771 4742 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924780 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924788 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924797 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924807 4742 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924815 4742 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924826 4742 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924836 4742 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924845 4742 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924853 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924863 4742 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924871 4742 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924879 4742 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924889 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924897 4742 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924909 4742 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924917 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924925 4742 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924934 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924943 4742 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924951 4742 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924960 4742 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924968 4742 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924976 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924985 4742 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.924993 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.925001 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.925009 4742 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.925017 4742 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.925029 4742 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.925036 4742 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.925044 4742 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.925070 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.925079 4742 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.925088 4742 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.925096 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.926161 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.927957 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.928852 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.930343 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.931425 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.931922 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.933307 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.933772 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.934547 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.935324 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.936505 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.937084 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.938165 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.939010 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.939954 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.940450 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.940933 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.941942 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.942812 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.943376 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.943798 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.954744 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.965105 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.980261 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.982212 4742 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610" exitCode=255 Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.982266 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610"} Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.987005 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:10:54 crc kubenswrapper[4742]: E1004 03:10:54.991094 4742 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:10:54 crc kubenswrapper[4742]: I1004 03:10:54.991445 4742 scope.go:117] "RemoveContainer" containerID="8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610" Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.001839 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.020445 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.042753 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.059713 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.073487 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.097529 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.101041 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.109398 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.109673 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.117157 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:10:55 crc kubenswrapper[4742]: W1004 03:10:55.122428 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-9e574f9119400085c516abb16a8e6e5332de65f94e0f35106cf6cd4887f3e909 WatchSource:0}: Error finding container 9e574f9119400085c516abb16a8e6e5332de65f94e0f35106cf6cd4887f3e909: Status 404 returned error can't find the container with id 9e574f9119400085c516abb16a8e6e5332de65f94e0f35106cf6cd4887f3e909 Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.123016 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.123095 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.125295 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.125894 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.140127 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.155570 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:10:55 crc kubenswrapper[4742]: W1004 03:10:55.160644 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-ed9821a16409b3c9f735591866e9c078faa788a45d36c9847468c65847b292ba WatchSource:0}: Error finding container ed9821a16409b3c9f735591866e9c078faa788a45d36c9847468c65847b292ba: Status 404 returned error can't find the container with id ed9821a16409b3c9f735591866e9c078faa788a45d36c9847468c65847b292ba Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.175577 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.190242 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.201352 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.212426 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.224603 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.234891 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.246075 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.258164 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.328226 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:10:55 crc kubenswrapper[4742]: E1004 03:10:55.328426 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:10:56.328394387 +0000 UTC m=+22.260731563 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.328527 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:10:55 crc kubenswrapper[4742]: E1004 03:10:55.328666 4742 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:10:55 crc kubenswrapper[4742]: E1004 03:10:55.328795 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:10:56.328753636 +0000 UTC m=+22.261090802 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.429286 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.429374 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.429408 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:10:55 crc kubenswrapper[4742]: E1004 03:10:55.429420 4742 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:10:55 crc kubenswrapper[4742]: E1004 03:10:55.429440 4742 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:10:55 crc kubenswrapper[4742]: E1004 03:10:55.429450 4742 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:10:55 crc kubenswrapper[4742]: E1004 03:10:55.429488 4742 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:10:55 crc kubenswrapper[4742]: E1004 03:10:55.429505 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 03:10:56.429490813 +0000 UTC m=+22.361827989 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:10:55 crc kubenswrapper[4742]: E1004 03:10:55.429523 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:10:56.429513443 +0000 UTC m=+22.361850619 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:10:55 crc kubenswrapper[4742]: E1004 03:10:55.429596 4742 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:10:55 crc kubenswrapper[4742]: E1004 03:10:55.429627 4742 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:10:55 crc kubenswrapper[4742]: E1004 03:10:55.429640 4742 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:10:55 crc kubenswrapper[4742]: E1004 03:10:55.429711 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 03:10:56.429691008 +0000 UTC m=+22.362028264 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.981826 4742 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.986004 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.987298 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8"} Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.987593 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.988392 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"acebe5f6cfbea325979af6336c871d1a6b38e6967a35cfc3f685f16e8f9159f3"} Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.989833 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2"} Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.989865 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01"} Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.989880 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ed9821a16409b3c9f735591866e9c078faa788a45d36c9847468c65847b292ba"} Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.991529 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc"} Oct 04 03:10:55 crc kubenswrapper[4742]: I1004 03:10:55.991553 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"9e574f9119400085c516abb16a8e6e5332de65f94e0f35106cf6cd4887f3e909"} Oct 04 03:10:56 crc kubenswrapper[4742]: I1004 03:10:56.004157 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:56 crc kubenswrapper[4742]: I1004 03:10:56.015391 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:56 crc kubenswrapper[4742]: I1004 03:10:56.025682 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:56 crc kubenswrapper[4742]: I1004 03:10:56.043523 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:56 crc kubenswrapper[4742]: I1004 03:10:56.066416 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:56 crc kubenswrapper[4742]: I1004 03:10:56.081436 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:56 crc kubenswrapper[4742]: I1004 03:10:56.093989 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:56 crc kubenswrapper[4742]: I1004 03:10:56.107097 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:56 crc kubenswrapper[4742]: I1004 03:10:56.120123 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:56 crc kubenswrapper[4742]: I1004 03:10:56.133808 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:56 crc kubenswrapper[4742]: I1004 03:10:56.150785 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:56 crc kubenswrapper[4742]: I1004 03:10:56.164241 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:56 crc kubenswrapper[4742]: I1004 03:10:56.179150 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:56 crc kubenswrapper[4742]: I1004 03:10:56.203315 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:56 crc kubenswrapper[4742]: I1004 03:10:56.237560 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:56 crc kubenswrapper[4742]: I1004 03:10:56.263472 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:56 crc kubenswrapper[4742]: I1004 03:10:56.336527 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:10:56 crc kubenswrapper[4742]: I1004 03:10:56.336605 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:10:56 crc kubenswrapper[4742]: E1004 03:10:56.336707 4742 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:10:56 crc kubenswrapper[4742]: E1004 03:10:56.336760 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:10:58.336726764 +0000 UTC m=+24.269063950 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:10:56 crc kubenswrapper[4742]: E1004 03:10:56.336819 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:10:58.336808636 +0000 UTC m=+24.269145932 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:10:56 crc kubenswrapper[4742]: I1004 03:10:56.437593 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:10:56 crc kubenswrapper[4742]: I1004 03:10:56.437633 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:10:56 crc kubenswrapper[4742]: I1004 03:10:56.437661 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:10:56 crc kubenswrapper[4742]: E1004 03:10:56.437765 4742 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:10:56 crc kubenswrapper[4742]: E1004 03:10:56.437843 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:10:58.43782516 +0000 UTC m=+24.370162336 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:10:56 crc kubenswrapper[4742]: E1004 03:10:56.437770 4742 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:10:56 crc kubenswrapper[4742]: E1004 03:10:56.437877 4742 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:10:56 crc kubenswrapper[4742]: E1004 03:10:56.437916 4742 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:10:56 crc kubenswrapper[4742]: E1004 03:10:56.437939 4742 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:10:56 crc kubenswrapper[4742]: E1004 03:10:56.437896 4742 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:10:56 crc kubenswrapper[4742]: E1004 03:10:56.437967 4742 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:10:56 crc kubenswrapper[4742]: E1004 03:10:56.437995 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 03:10:58.437975634 +0000 UTC m=+24.370312810 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:10:56 crc kubenswrapper[4742]: E1004 03:10:56.438021 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 03:10:58.438006745 +0000 UTC m=+24.370343981 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:10:56 crc kubenswrapper[4742]: I1004 03:10:56.857977 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:10:56 crc kubenswrapper[4742]: I1004 03:10:56.858037 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:10:56 crc kubenswrapper[4742]: I1004 03:10:56.858060 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:10:56 crc kubenswrapper[4742]: E1004 03:10:56.858121 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:10:56 crc kubenswrapper[4742]: E1004 03:10:56.858236 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:10:56 crc kubenswrapper[4742]: E1004 03:10:56.858412 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:10:57 crc kubenswrapper[4742]: I1004 03:10:57.708326 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-shjbc"] Oct 04 03:10:57 crc kubenswrapper[4742]: I1004 03:10:57.708619 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-shjbc" Oct 04 03:10:57 crc kubenswrapper[4742]: I1004 03:10:57.713355 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 04 03:10:57 crc kubenswrapper[4742]: I1004 03:10:57.715170 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 04 03:10:57 crc kubenswrapper[4742]: I1004 03:10:57.715725 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 04 03:10:57 crc kubenswrapper[4742]: I1004 03:10:57.745411 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:57 crc kubenswrapper[4742]: I1004 03:10:57.748475 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/b8e4f883-06de-41e5-83c1-90f55f41061c-hosts-file\") pod \"node-resolver-shjbc\" (UID: \"b8e4f883-06de-41e5-83c1-90f55f41061c\") " pod="openshift-dns/node-resolver-shjbc" Oct 04 03:10:57 crc kubenswrapper[4742]: I1004 03:10:57.748516 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8xlx\" (UniqueName: \"kubernetes.io/projected/b8e4f883-06de-41e5-83c1-90f55f41061c-kube-api-access-s8xlx\") pod \"node-resolver-shjbc\" (UID: \"b8e4f883-06de-41e5-83c1-90f55f41061c\") " pod="openshift-dns/node-resolver-shjbc" Oct 04 03:10:57 crc kubenswrapper[4742]: I1004 03:10:57.772128 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:57 crc kubenswrapper[4742]: I1004 03:10:57.787985 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:57 crc kubenswrapper[4742]: I1004 03:10:57.802126 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:57 crc kubenswrapper[4742]: I1004 03:10:57.814412 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:57 crc kubenswrapper[4742]: I1004 03:10:57.834938 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:57 crc kubenswrapper[4742]: I1004 03:10:57.849477 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8xlx\" (UniqueName: \"kubernetes.io/projected/b8e4f883-06de-41e5-83c1-90f55f41061c-kube-api-access-s8xlx\") pod \"node-resolver-shjbc\" (UID: \"b8e4f883-06de-41e5-83c1-90f55f41061c\") " pod="openshift-dns/node-resolver-shjbc" Oct 04 03:10:57 crc kubenswrapper[4742]: I1004 03:10:57.849562 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/b8e4f883-06de-41e5-83c1-90f55f41061c-hosts-file\") pod \"node-resolver-shjbc\" (UID: \"b8e4f883-06de-41e5-83c1-90f55f41061c\") " pod="openshift-dns/node-resolver-shjbc" Oct 04 03:10:57 crc kubenswrapper[4742]: I1004 03:10:57.849735 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/b8e4f883-06de-41e5-83c1-90f55f41061c-hosts-file\") pod \"node-resolver-shjbc\" (UID: \"b8e4f883-06de-41e5-83c1-90f55f41061c\") " pod="openshift-dns/node-resolver-shjbc" Oct 04 03:10:57 crc kubenswrapper[4742]: I1004 03:10:57.852216 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:57 crc kubenswrapper[4742]: I1004 03:10:57.873646 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8xlx\" (UniqueName: \"kubernetes.io/projected/b8e4f883-06de-41e5-83c1-90f55f41061c-kube-api-access-s8xlx\") pod \"node-resolver-shjbc\" (UID: \"b8e4f883-06de-41e5-83c1-90f55f41061c\") " pod="openshift-dns/node-resolver-shjbc" Oct 04 03:10:57 crc kubenswrapper[4742]: I1004 03:10:57.882993 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:57 crc kubenswrapper[4742]: I1004 03:10:57.895490 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.004158 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1"} Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.021408 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-shjbc" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.022631 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: W1004 03:10:58.032567 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb8e4f883_06de_41e5_83c1_90f55f41061c.slice/crio-65fe4ba17b14ce316ed9152916268d073b422230933ea84b9f24bfc2488b7c44 WatchSource:0}: Error finding container 65fe4ba17b14ce316ed9152916268d073b422230933ea84b9f24bfc2488b7c44: Status 404 returned error can't find the container with id 65fe4ba17b14ce316ed9152916268d073b422230933ea84b9f24bfc2488b7c44 Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.044163 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.071027 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.087601 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.108654 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.121035 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-qdt7v"] Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.121479 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.123466 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.123575 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.123470 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.124253 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.124595 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.126070 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.140855 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.154068 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.174573 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.207028 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.241530 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.253380 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/989ea247-7ac9-4ade-8dcb-2632fef4f83e-rootfs\") pod \"machine-config-daemon-qdt7v\" (UID: \"989ea247-7ac9-4ade-8dcb-2632fef4f83e\") " pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.253441 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/989ea247-7ac9-4ade-8dcb-2632fef4f83e-mcd-auth-proxy-config\") pod \"machine-config-daemon-qdt7v\" (UID: \"989ea247-7ac9-4ade-8dcb-2632fef4f83e\") " pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.253467 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/989ea247-7ac9-4ade-8dcb-2632fef4f83e-proxy-tls\") pod \"machine-config-daemon-qdt7v\" (UID: \"989ea247-7ac9-4ade-8dcb-2632fef4f83e\") " pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.253493 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8s6xh\" (UniqueName: \"kubernetes.io/projected/989ea247-7ac9-4ade-8dcb-2632fef4f83e-kube-api-access-8s6xh\") pod \"machine-config-daemon-qdt7v\" (UID: \"989ea247-7ac9-4ade-8dcb-2632fef4f83e\") " pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.279896 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.318672 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.343314 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.354047 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.354124 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/989ea247-7ac9-4ade-8dcb-2632fef4f83e-mcd-auth-proxy-config\") pod \"machine-config-daemon-qdt7v\" (UID: \"989ea247-7ac9-4ade-8dcb-2632fef4f83e\") " pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.354147 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/989ea247-7ac9-4ade-8dcb-2632fef4f83e-proxy-tls\") pod \"machine-config-daemon-qdt7v\" (UID: \"989ea247-7ac9-4ade-8dcb-2632fef4f83e\") " pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.354179 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8s6xh\" (UniqueName: \"kubernetes.io/projected/989ea247-7ac9-4ade-8dcb-2632fef4f83e-kube-api-access-8s6xh\") pod \"machine-config-daemon-qdt7v\" (UID: \"989ea247-7ac9-4ade-8dcb-2632fef4f83e\") " pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.354208 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/989ea247-7ac9-4ade-8dcb-2632fef4f83e-rootfs\") pod \"machine-config-daemon-qdt7v\" (UID: \"989ea247-7ac9-4ade-8dcb-2632fef4f83e\") " pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.354228 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:10:58 crc kubenswrapper[4742]: E1004 03:10:58.354331 4742 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:10:58 crc kubenswrapper[4742]: E1004 03:10:58.354375 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:11:02.354363393 +0000 UTC m=+28.286700559 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:10:58 crc kubenswrapper[4742]: E1004 03:10:58.354619 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:11:02.35460921 +0000 UTC m=+28.286946386 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.355611 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/989ea247-7ac9-4ade-8dcb-2632fef4f83e-mcd-auth-proxy-config\") pod \"machine-config-daemon-qdt7v\" (UID: \"989ea247-7ac9-4ade-8dcb-2632fef4f83e\") " pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.355667 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/989ea247-7ac9-4ade-8dcb-2632fef4f83e-rootfs\") pod \"machine-config-daemon-qdt7v\" (UID: \"989ea247-7ac9-4ade-8dcb-2632fef4f83e\") " pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.360093 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/989ea247-7ac9-4ade-8dcb-2632fef4f83e-proxy-tls\") pod \"machine-config-daemon-qdt7v\" (UID: \"989ea247-7ac9-4ade-8dcb-2632fef4f83e\") " pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.363628 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.371473 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8s6xh\" (UniqueName: \"kubernetes.io/projected/989ea247-7ac9-4ade-8dcb-2632fef4f83e-kube-api-access-8s6xh\") pod \"machine-config-daemon-qdt7v\" (UID: \"989ea247-7ac9-4ade-8dcb-2632fef4f83e\") " pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.388340 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.405331 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.420733 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.432869 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.434492 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: W1004 03:10:58.443278 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod989ea247_7ac9_4ade_8dcb_2632fef4f83e.slice/crio-da2fe175b5a14f7f2f87d57739452b0e4a51701af529b477368d041689a5b1ae WatchSource:0}: Error finding container da2fe175b5a14f7f2f87d57739452b0e4a51701af529b477368d041689a5b1ae: Status 404 returned error can't find the container with id da2fe175b5a14f7f2f87d57739452b0e4a51701af529b477368d041689a5b1ae Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.455456 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.455506 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.455529 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:10:58 crc kubenswrapper[4742]: E1004 03:10:58.455675 4742 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:10:58 crc kubenswrapper[4742]: E1004 03:10:58.455692 4742 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:10:58 crc kubenswrapper[4742]: E1004 03:10:58.455704 4742 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:10:58 crc kubenswrapper[4742]: E1004 03:10:58.455760 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 03:11:02.455742526 +0000 UTC m=+28.388079692 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:10:58 crc kubenswrapper[4742]: E1004 03:10:58.455855 4742 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:10:58 crc kubenswrapper[4742]: E1004 03:10:58.455888 4742 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:10:58 crc kubenswrapper[4742]: E1004 03:10:58.455901 4742 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:10:58 crc kubenswrapper[4742]: E1004 03:10:58.455916 4742 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:10:58 crc kubenswrapper[4742]: E1004 03:10:58.455969 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 03:11:02.455949542 +0000 UTC m=+28.388286718 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:10:58 crc kubenswrapper[4742]: E1004 03:10:58.456054 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:11:02.456026294 +0000 UTC m=+28.388363470 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.527129 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-ll99m"] Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.528611 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-ll99m" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.530105 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-nc8m7"] Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.530585 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.532343 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.532358 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.532425 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.532366 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.532777 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.532978 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.536310 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.549540 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.563470 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.576006 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.589069 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.604315 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.619014 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.629660 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.644032 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.657794 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-host-var-lib-cni-multus\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.657838 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-host-var-lib-kubelet\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.657859 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-host-run-multus-certs\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.657876 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-os-release\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.657897 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pj2t\" (UniqueName: \"kubernetes.io/projected/46de52ee-226c-4713-bd7b-961f8f43f635-kube-api-access-8pj2t\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.657942 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/87c84d43-0375-42ca-b3b3-9cadda96fcbd-tuning-conf-dir\") pod \"multus-additional-cni-plugins-ll99m\" (UID: \"87c84d43-0375-42ca-b3b3-9cadda96fcbd\") " pod="openshift-multus/multus-additional-cni-plugins-ll99m" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.657966 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/87c84d43-0375-42ca-b3b3-9cadda96fcbd-system-cni-dir\") pod \"multus-additional-cni-plugins-ll99m\" (UID: \"87c84d43-0375-42ca-b3b3-9cadda96fcbd\") " pod="openshift-multus/multus-additional-cni-plugins-ll99m" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.658117 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-system-cni-dir\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.658294 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/87c84d43-0375-42ca-b3b3-9cadda96fcbd-cnibin\") pod \"multus-additional-cni-plugins-ll99m\" (UID: \"87c84d43-0375-42ca-b3b3-9cadda96fcbd\") " pod="openshift-multus/multus-additional-cni-plugins-ll99m" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.658343 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/87c84d43-0375-42ca-b3b3-9cadda96fcbd-cni-binary-copy\") pod \"multus-additional-cni-plugins-ll99m\" (UID: \"87c84d43-0375-42ca-b3b3-9cadda96fcbd\") " pod="openshift-multus/multus-additional-cni-plugins-ll99m" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.658375 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-host-run-netns\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.658421 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-cnibin\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.658478 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/87c84d43-0375-42ca-b3b3-9cadda96fcbd-os-release\") pod \"multus-additional-cni-plugins-ll99m\" (UID: \"87c84d43-0375-42ca-b3b3-9cadda96fcbd\") " pod="openshift-multus/multus-additional-cni-plugins-ll99m" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.658528 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-host-run-k8s-cni-cncf-io\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.658663 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-hostroot\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.658749 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-etc-kubernetes\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.658802 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/87c84d43-0375-42ca-b3b3-9cadda96fcbd-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-ll99m\" (UID: \"87c84d43-0375-42ca-b3b3-9cadda96fcbd\") " pod="openshift-multus/multus-additional-cni-plugins-ll99m" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.658842 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-multus-conf-dir\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.658875 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/46de52ee-226c-4713-bd7b-961f8f43f635-multus-daemon-config\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.658905 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-multus-socket-dir-parent\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.658931 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-host-var-lib-cni-bin\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.658963 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dgsv\" (UniqueName: \"kubernetes.io/projected/87c84d43-0375-42ca-b3b3-9cadda96fcbd-kube-api-access-9dgsv\") pod \"multus-additional-cni-plugins-ll99m\" (UID: \"87c84d43-0375-42ca-b3b3-9cadda96fcbd\") " pod="openshift-multus/multus-additional-cni-plugins-ll99m" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.658994 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-multus-cni-dir\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.659151 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/46de52ee-226c-4713-bd7b-961f8f43f635-cni-binary-copy\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.659220 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.673100 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.687401 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.700098 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.720629 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.737007 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.752216 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.760165 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/87c84d43-0375-42ca-b3b3-9cadda96fcbd-tuning-conf-dir\") pod \"multus-additional-cni-plugins-ll99m\" (UID: \"87c84d43-0375-42ca-b3b3-9cadda96fcbd\") " pod="openshift-multus/multus-additional-cni-plugins-ll99m" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.760205 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/87c84d43-0375-42ca-b3b3-9cadda96fcbd-system-cni-dir\") pod \"multus-additional-cni-plugins-ll99m\" (UID: \"87c84d43-0375-42ca-b3b3-9cadda96fcbd\") " pod="openshift-multus/multus-additional-cni-plugins-ll99m" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.760223 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-system-cni-dir\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.760250 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/87c84d43-0375-42ca-b3b3-9cadda96fcbd-cnibin\") pod \"multus-additional-cni-plugins-ll99m\" (UID: \"87c84d43-0375-42ca-b3b3-9cadda96fcbd\") " pod="openshift-multus/multus-additional-cni-plugins-ll99m" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.760283 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/87c84d43-0375-42ca-b3b3-9cadda96fcbd-cni-binary-copy\") pod \"multus-additional-cni-plugins-ll99m\" (UID: \"87c84d43-0375-42ca-b3b3-9cadda96fcbd\") " pod="openshift-multus/multus-additional-cni-plugins-ll99m" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.760333 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/87c84d43-0375-42ca-b3b3-9cadda96fcbd-system-cni-dir\") pod \"multus-additional-cni-plugins-ll99m\" (UID: \"87c84d43-0375-42ca-b3b3-9cadda96fcbd\") " pod="openshift-multus/multus-additional-cni-plugins-ll99m" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.760392 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-host-run-netns\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.760301 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-host-run-netns\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.760405 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/87c84d43-0375-42ca-b3b3-9cadda96fcbd-cnibin\") pod \"multus-additional-cni-plugins-ll99m\" (UID: \"87c84d43-0375-42ca-b3b3-9cadda96fcbd\") " pod="openshift-multus/multus-additional-cni-plugins-ll99m" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.760514 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-cnibin\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.760416 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-system-cni-dir\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.760648 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-cnibin\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.760723 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/87c84d43-0375-42ca-b3b3-9cadda96fcbd-os-release\") pod \"multus-additional-cni-plugins-ll99m\" (UID: \"87c84d43-0375-42ca-b3b3-9cadda96fcbd\") " pod="openshift-multus/multus-additional-cni-plugins-ll99m" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.760802 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/87c84d43-0375-42ca-b3b3-9cadda96fcbd-tuning-conf-dir\") pod \"multus-additional-cni-plugins-ll99m\" (UID: \"87c84d43-0375-42ca-b3b3-9cadda96fcbd\") " pod="openshift-multus/multus-additional-cni-plugins-ll99m" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.760817 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-host-run-k8s-cni-cncf-io\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.760870 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-etc-kubernetes\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.760908 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-host-run-k8s-cni-cncf-io\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.760924 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-hostroot\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.760963 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-etc-kubernetes\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.760976 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/87c84d43-0375-42ca-b3b3-9cadda96fcbd-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-ll99m\" (UID: \"87c84d43-0375-42ca-b3b3-9cadda96fcbd\") " pod="openshift-multus/multus-additional-cni-plugins-ll99m" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.761038 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-multus-conf-dir\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.761072 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-multus-conf-dir\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.761087 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/46de52ee-226c-4713-bd7b-961f8f43f635-multus-daemon-config\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.761115 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/87c84d43-0375-42ca-b3b3-9cadda96fcbd-cni-binary-copy\") pod \"multus-additional-cni-plugins-ll99m\" (UID: \"87c84d43-0375-42ca-b3b3-9cadda96fcbd\") " pod="openshift-multus/multus-additional-cni-plugins-ll99m" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.761139 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-multus-socket-dir-parent\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.760969 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-hostroot\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.761129 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/87c84d43-0375-42ca-b3b3-9cadda96fcbd-os-release\") pod \"multus-additional-cni-plugins-ll99m\" (UID: \"87c84d43-0375-42ca-b3b3-9cadda96fcbd\") " pod="openshift-multus/multus-additional-cni-plugins-ll99m" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.761202 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-multus-socket-dir-parent\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.761200 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-host-var-lib-cni-bin\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.761323 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dgsv\" (UniqueName: \"kubernetes.io/projected/87c84d43-0375-42ca-b3b3-9cadda96fcbd-kube-api-access-9dgsv\") pod \"multus-additional-cni-plugins-ll99m\" (UID: \"87c84d43-0375-42ca-b3b3-9cadda96fcbd\") " pod="openshift-multus/multus-additional-cni-plugins-ll99m" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.761224 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-host-var-lib-cni-bin\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.761385 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-multus-cni-dir\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.761459 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/46de52ee-226c-4713-bd7b-961f8f43f635-cni-binary-copy\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.761505 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-os-release\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.761550 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-host-var-lib-cni-multus\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.761601 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-multus-cni-dir\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.761605 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-host-var-lib-kubelet\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.761605 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/87c84d43-0375-42ca-b3b3-9cadda96fcbd-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-ll99m\" (UID: \"87c84d43-0375-42ca-b3b3-9cadda96fcbd\") " pod="openshift-multus/multus-additional-cni-plugins-ll99m" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.761635 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-host-var-lib-cni-multus\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.761669 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-host-run-multus-certs\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.761692 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-os-release\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.761720 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pj2t\" (UniqueName: \"kubernetes.io/projected/46de52ee-226c-4713-bd7b-961f8f43f635-kube-api-access-8pj2t\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.761690 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-host-var-lib-kubelet\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.761741 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/46de52ee-226c-4713-bd7b-961f8f43f635-host-run-multus-certs\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.762134 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/46de52ee-226c-4713-bd7b-961f8f43f635-multus-daemon-config\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.762207 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/46de52ee-226c-4713-bd7b-961f8f43f635-cni-binary-copy\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.767183 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc8m7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46de52ee-226c-4713-bd7b-961f8f43f635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pj2t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc8m7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.779958 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dgsv\" (UniqueName: \"kubernetes.io/projected/87c84d43-0375-42ca-b3b3-9cadda96fcbd-kube-api-access-9dgsv\") pod \"multus-additional-cni-plugins-ll99m\" (UID: \"87c84d43-0375-42ca-b3b3-9cadda96fcbd\") " pod="openshift-multus/multus-additional-cni-plugins-ll99m" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.780667 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.780984 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pj2t\" (UniqueName: \"kubernetes.io/projected/46de52ee-226c-4713-bd7b-961f8f43f635-kube-api-access-8pj2t\") pod \"multus-nc8m7\" (UID: \"46de52ee-226c-4713-bd7b-961f8f43f635\") " pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.794316 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.806715 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.819351 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.830967 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.846251 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.847354 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-ll99m" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.853538 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-nc8m7" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.857887 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.857953 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:10:58 crc kubenswrapper[4742]: E1004 03:10:58.857985 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:10:58 crc kubenswrapper[4742]: E1004 03:10:58.858115 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.858292 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:10:58 crc kubenswrapper[4742]: E1004 03:10:58.858362 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:10:58 crc kubenswrapper[4742]: W1004 03:10:58.859950 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod87c84d43_0375_42ca_b3b3_9cadda96fcbd.slice/crio-6cef787c28a7d62d907f6240203ae2cc80729a1eb0b8bfbb1c8a1ba1a46223c6 WatchSource:0}: Error finding container 6cef787c28a7d62d907f6240203ae2cc80729a1eb0b8bfbb1c8a1ba1a46223c6: Status 404 returned error can't find the container with id 6cef787c28a7d62d907f6240203ae2cc80729a1eb0b8bfbb1c8a1ba1a46223c6 Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.861140 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: W1004 03:10:58.865126 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46de52ee_226c_4713_bd7b_961f8f43f635.slice/crio-e0aba64e9fd6899ba87e54bd5a5d7a548db0af1c4bef83a42a0ca3c558dbc4b3 WatchSource:0}: Error finding container e0aba64e9fd6899ba87e54bd5a5d7a548db0af1c4bef83a42a0ca3c558dbc4b3: Status 404 returned error can't find the container with id e0aba64e9fd6899ba87e54bd5a5d7a548db0af1c4bef83a42a0ca3c558dbc4b3 Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.891151 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9qrtl"] Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.892190 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.894474 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.895042 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.895620 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.895635 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.895673 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.895802 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.895819 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.913189 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.955440 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.962686 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-run-netns\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.962825 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.962907 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-node-log\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.962997 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-log-socket\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.963065 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-run-openvswitch\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.963154 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-systemd-units\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.963225 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-cni-netd\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.963310 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-env-overrides\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.963402 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-cni-bin\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.963875 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-ovnkube-script-lib\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.963969 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-kubelet\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.964054 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txjzp\" (UniqueName: \"kubernetes.io/projected/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-kube-api-access-txjzp\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.964139 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-slash\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.964217 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-run-ovn-kubernetes\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.964308 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-run-systemd\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.964381 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-var-lib-openvswitch\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.964453 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-ovnkube-config\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.964524 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-run-ovn\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.964605 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-etc-openvswitch\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:58 crc kubenswrapper[4742]: I1004 03:10:58.964685 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-ovn-node-metrics-cert\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.008183 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-shjbc" event={"ID":"b8e4f883-06de-41e5-83c1-90f55f41061c","Type":"ContainerStarted","Data":"e11b178c3523269e8b4c09a1ad6f86eaee1d8f19f5056840dfa24466248889a4"} Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.008227 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-shjbc" event={"ID":"b8e4f883-06de-41e5-83c1-90f55f41061c","Type":"ContainerStarted","Data":"65fe4ba17b14ce316ed9152916268d073b422230933ea84b9f24bfc2488b7c44"} Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.013903 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" event={"ID":"87c84d43-0375-42ca-b3b3-9cadda96fcbd","Type":"ContainerStarted","Data":"6cef787c28a7d62d907f6240203ae2cc80729a1eb0b8bfbb1c8a1ba1a46223c6"} Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.015077 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nc8m7" event={"ID":"46de52ee-226c-4713-bd7b-961f8f43f635","Type":"ContainerStarted","Data":"e0aba64e9fd6899ba87e54bd5a5d7a548db0af1c4bef83a42a0ca3c558dbc4b3"} Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.016930 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" event={"ID":"989ea247-7ac9-4ade-8dcb-2632fef4f83e","Type":"ContainerStarted","Data":"261e8ace3d214aec7b737ea538ac0996cc11a5dae65eacc33dd3a3ea65ed3558"} Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.016956 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" event={"ID":"989ea247-7ac9-4ade-8dcb-2632fef4f83e","Type":"ContainerStarted","Data":"dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde"} Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.016965 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" event={"ID":"989ea247-7ac9-4ade-8dcb-2632fef4f83e","Type":"ContainerStarted","Data":"da2fe175b5a14f7f2f87d57739452b0e4a51701af529b477368d041689a5b1ae"} Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.018236 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.059633 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.065516 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-var-lib-openvswitch\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.065561 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-run-systemd\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.065578 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-ovnkube-config\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.065594 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-run-ovn\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.065612 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-etc-openvswitch\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.065627 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-ovn-node-metrics-cert\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.065649 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.065667 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-run-netns\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.065680 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-node-log\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.065697 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-log-socket\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.065709 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-run-openvswitch\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.065760 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-systemd-units\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.065784 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-cni-netd\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.065811 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-env-overrides\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.065831 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-cni-bin\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.065848 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-ovnkube-script-lib\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.065854 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-run-netns\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.065916 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-kubelet\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.065921 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-node-log\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.065948 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-var-lib-openvswitch\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.065960 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-log-socket\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.065983 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-run-systemd\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.066006 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-run-openvswitch\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.066690 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-ovnkube-config\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.066693 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-run-ovn\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.066791 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-etc-openvswitch\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.066912 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-systemd-units\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.067124 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.067156 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-cni-netd\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.065868 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-kubelet\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.067183 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txjzp\" (UniqueName: \"kubernetes.io/projected/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-kube-api-access-txjzp\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.067209 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-slash\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.067233 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-run-ovn-kubernetes\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.067246 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-env-overrides\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.067285 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-run-ovn-kubernetes\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.067307 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-cni-bin\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.067318 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-slash\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.067862 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-ovnkube-script-lib\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.070508 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-ovn-node-metrics-cert\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.087643 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc8m7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46de52ee-226c-4713-bd7b-961f8f43f635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pj2t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc8m7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.096003 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txjzp\" (UniqueName: \"kubernetes.io/projected/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-kube-api-access-txjzp\") pod \"ovnkube-node-9qrtl\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.115459 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.132130 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.142708 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.161081 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qrtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.173490 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.187445 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.201188 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.206012 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.212473 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:59 crc kubenswrapper[4742]: W1004 03:10:59.219366 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0cfc6e64_ba08_4c44_a18e_3b93ab792a6a.slice/crio-fbd3b1a032009427bfc30386ecd659463d8b4e86e8278e906adb3c7f353e9e91 WatchSource:0}: Error finding container fbd3b1a032009427bfc30386ecd659463d8b4e86e8278e906adb3c7f353e9e91: Status 404 returned error can't find the container with id fbd3b1a032009427bfc30386ecd659463d8b4e86e8278e906adb3c7f353e9e91 Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.230528 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://261e8ace3d214aec7b737ea538ac0996cc11a5dae65eacc33dd3a3ea65ed3558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.249747 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.264099 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.277737 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11b178c3523269e8b4c09a1ad6f86eaee1d8f19f5056840dfa24466248889a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.291820 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.304070 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc8m7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46de52ee-226c-4713-bd7b-961f8f43f635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pj2t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc8m7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.318323 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.332714 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.347403 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.360134 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.378044 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qrtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.389748 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:10:59 crc kubenswrapper[4742]: I1004 03:10:59.403026 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:10:59Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.023044 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nc8m7" event={"ID":"46de52ee-226c-4713-bd7b-961f8f43f635","Type":"ContainerStarted","Data":"7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c"} Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.026095 4742 generic.go:334] "Generic (PLEG): container finished" podID="87c84d43-0375-42ca-b3b3-9cadda96fcbd" containerID="724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f" exitCode=0 Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.026158 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" event={"ID":"87c84d43-0375-42ca-b3b3-9cadda96fcbd","Type":"ContainerDied","Data":"724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f"} Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.028613 4742 generic.go:334] "Generic (PLEG): container finished" podID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerID="53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9" exitCode=0 Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.028637 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" event={"ID":"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a","Type":"ContainerDied","Data":"53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9"} Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.028676 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" event={"ID":"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a","Type":"ContainerStarted","Data":"fbd3b1a032009427bfc30386ecd659463d8b4e86e8278e906adb3c7f353e9e91"} Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.046486 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.076254 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc8m7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46de52ee-226c-4713-bd7b-961f8f43f635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pj2t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc8m7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.095027 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.110301 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.124788 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.146773 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.165997 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qrtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.180413 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.194079 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.207990 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://261e8ace3d214aec7b737ea538ac0996cc11a5dae65eacc33dd3a3ea65ed3558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.223898 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.243212 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.257021 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11b178c3523269e8b4c09a1ad6f86eaee1d8f19f5056840dfa24466248889a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.270042 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.280072 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11b178c3523269e8b4c09a1ad6f86eaee1d8f19f5056840dfa24466248889a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.294734 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.312005 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.328482 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.342464 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.357107 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc8m7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46de52ee-226c-4713-bd7b-961f8f43f635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pj2t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc8m7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.374738 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.390749 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.406022 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.406237 4742 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.408177 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.408230 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.408238 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.408388 4742 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.417069 4742 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.417523 4742 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.418894 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.418944 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.418956 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.418971 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.418982 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:00Z","lastTransitionTime":"2025-10-04T03:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.425766 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qrtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.438020 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://261e8ace3d214aec7b737ea538ac0996cc11a5dae65eacc33dd3a3ea65ed3558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:00 crc kubenswrapper[4742]: E1004 03:11:00.439978 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.443210 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.443263 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.443291 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.443350 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.443367 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:00Z","lastTransitionTime":"2025-10-04T03:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.452183 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:00 crc kubenswrapper[4742]: E1004 03:11:00.457438 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.461532 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.461584 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.461598 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.461616 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.461634 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:00Z","lastTransitionTime":"2025-10-04T03:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:00 crc kubenswrapper[4742]: E1004 03:11:00.473617 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.482143 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.482181 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.482194 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.482216 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.482231 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:00Z","lastTransitionTime":"2025-10-04T03:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:00 crc kubenswrapper[4742]: E1004 03:11:00.494368 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.498591 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.498635 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.498647 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.498665 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.498677 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:00Z","lastTransitionTime":"2025-10-04T03:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:00 crc kubenswrapper[4742]: E1004 03:11:00.511767 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:00 crc kubenswrapper[4742]: E1004 03:11:00.511907 4742 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.514027 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.514090 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.514104 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.514123 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.514134 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:00Z","lastTransitionTime":"2025-10-04T03:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.616101 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.616560 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.616571 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.616595 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.616623 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:00Z","lastTransitionTime":"2025-10-04T03:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.718381 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.718424 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.718439 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.718457 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.718469 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:00Z","lastTransitionTime":"2025-10-04T03:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.822017 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.822075 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.822091 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.822114 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.822126 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:00Z","lastTransitionTime":"2025-10-04T03:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.857582 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.857580 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.857766 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:11:00 crc kubenswrapper[4742]: E1004 03:11:00.857911 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:11:00 crc kubenswrapper[4742]: E1004 03:11:00.858155 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:11:00 crc kubenswrapper[4742]: E1004 03:11:00.858439 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.925633 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.925679 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.925688 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.925705 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:00 crc kubenswrapper[4742]: I1004 03:11:00.925716 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:00Z","lastTransitionTime":"2025-10-04T03:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.028570 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.028614 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.028628 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.028647 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.028659 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:01Z","lastTransitionTime":"2025-10-04T03:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.033091 4742 generic.go:334] "Generic (PLEG): container finished" podID="87c84d43-0375-42ca-b3b3-9cadda96fcbd" containerID="9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e" exitCode=0 Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.033154 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" event={"ID":"87c84d43-0375-42ca-b3b3-9cadda96fcbd","Type":"ContainerDied","Data":"9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e"} Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.039516 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" event={"ID":"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a","Type":"ContainerStarted","Data":"0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433"} Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.039572 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" event={"ID":"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a","Type":"ContainerStarted","Data":"e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6"} Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.039586 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" event={"ID":"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a","Type":"ContainerStarted","Data":"ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236"} Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.039624 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" event={"ID":"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a","Type":"ContainerStarted","Data":"ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993"} Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.039638 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" event={"ID":"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a","Type":"ContainerStarted","Data":"73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b"} Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.039652 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" event={"ID":"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a","Type":"ContainerStarted","Data":"78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24"} Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.050972 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.071106 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.087782 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.101536 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.116178 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc8m7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46de52ee-226c-4713-bd7b-961f8f43f635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pj2t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc8m7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.130571 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.131876 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.131919 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.131932 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.131962 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.131988 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-cwm6r"] Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.131974 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:01Z","lastTransitionTime":"2025-10-04T03:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.132587 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-cwm6r" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.135887 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.135974 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.137471 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.137729 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.146402 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.158845 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.177124 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qrtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.188468 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://261e8ace3d214aec7b737ea538ac0996cc11a5dae65eacc33dd3a3ea65ed3558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.191443 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/09a5aee9-b917-473a-9196-5cde6553c72a-serviceca\") pod \"node-ca-cwm6r\" (UID: \"09a5aee9-b917-473a-9196-5cde6553c72a\") " pod="openshift-image-registry/node-ca-cwm6r" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.191489 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09a5aee9-b917-473a-9196-5cde6553c72a-host\") pod \"node-ca-cwm6r\" (UID: \"09a5aee9-b917-473a-9196-5cde6553c72a\") " pod="openshift-image-registry/node-ca-cwm6r" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.191518 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrdmr\" (UniqueName: \"kubernetes.io/projected/09a5aee9-b917-473a-9196-5cde6553c72a-kube-api-access-zrdmr\") pod \"node-ca-cwm6r\" (UID: \"09a5aee9-b917-473a-9196-5cde6553c72a\") " pod="openshift-image-registry/node-ca-cwm6r" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.204621 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.220863 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.233699 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11b178c3523269e8b4c09a1ad6f86eaee1d8f19f5056840dfa24466248889a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.234995 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.235040 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.235050 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.235067 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.235078 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:01Z","lastTransitionTime":"2025-10-04T03:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.247730 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://261e8ace3d214aec7b737ea538ac0996cc11a5dae65eacc33dd3a3ea65ed3558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.264823 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.279223 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cwm6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a5aee9-b917-473a-9196-5cde6553c72a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrdmr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cwm6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.293097 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/09a5aee9-b917-473a-9196-5cde6553c72a-serviceca\") pod \"node-ca-cwm6r\" (UID: \"09a5aee9-b917-473a-9196-5cde6553c72a\") " pod="openshift-image-registry/node-ca-cwm6r" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.293157 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09a5aee9-b917-473a-9196-5cde6553c72a-host\") pod \"node-ca-cwm6r\" (UID: \"09a5aee9-b917-473a-9196-5cde6553c72a\") " pod="openshift-image-registry/node-ca-cwm6r" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.293200 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrdmr\" (UniqueName: \"kubernetes.io/projected/09a5aee9-b917-473a-9196-5cde6553c72a-kube-api-access-zrdmr\") pod \"node-ca-cwm6r\" (UID: \"09a5aee9-b917-473a-9196-5cde6553c72a\") " pod="openshift-image-registry/node-ca-cwm6r" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.293343 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09a5aee9-b917-473a-9196-5cde6553c72a-host\") pod \"node-ca-cwm6r\" (UID: \"09a5aee9-b917-473a-9196-5cde6553c72a\") " pod="openshift-image-registry/node-ca-cwm6r" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.294643 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/09a5aee9-b917-473a-9196-5cde6553c72a-serviceca\") pod \"node-ca-cwm6r\" (UID: \"09a5aee9-b917-473a-9196-5cde6553c72a\") " pod="openshift-image-registry/node-ca-cwm6r" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.295114 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.307730 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11b178c3523269e8b4c09a1ad6f86eaee1d8f19f5056840dfa24466248889a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.314016 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrdmr\" (UniqueName: \"kubernetes.io/projected/09a5aee9-b917-473a-9196-5cde6553c72a-kube-api-access-zrdmr\") pod \"node-ca-cwm6r\" (UID: \"09a5aee9-b917-473a-9196-5cde6553c72a\") " pod="openshift-image-registry/node-ca-cwm6r" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.324163 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc8m7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46de52ee-226c-4713-bd7b-961f8f43f635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pj2t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc8m7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.337455 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.337502 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.337515 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.337535 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.337548 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:01Z","lastTransitionTime":"2025-10-04T03:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.344545 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.361640 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.401514 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.439032 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.440763 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.440806 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.440819 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.440838 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.440849 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:01Z","lastTransitionTime":"2025-10-04T03:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.454680 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-cwm6r" Oct 04 03:11:01 crc kubenswrapper[4742]: W1004 03:11:01.466983 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09a5aee9_b917_473a_9196_5cde6553c72a.slice/crio-9688d82ad1672c2fd0714b838dc0a5d5a7a8381e9af453734a0d0894fb43ef78 WatchSource:0}: Error finding container 9688d82ad1672c2fd0714b838dc0a5d5a7a8381e9af453734a0d0894fb43ef78: Status 404 returned error can't find the container with id 9688d82ad1672c2fd0714b838dc0a5d5a7a8381e9af453734a0d0894fb43ef78 Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.490375 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qrtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.521519 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.543288 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.543327 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.543336 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.543349 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.543359 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:01Z","lastTransitionTime":"2025-10-04T03:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.562011 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.604745 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.646785 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.646909 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.646921 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.646940 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.646951 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:01Z","lastTransitionTime":"2025-10-04T03:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.750053 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.750115 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.750131 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.750154 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.750171 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:01Z","lastTransitionTime":"2025-10-04T03:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.853700 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.853770 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.853793 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.853824 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.853844 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:01Z","lastTransitionTime":"2025-10-04T03:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.956193 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.956245 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.956259 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.956300 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:01 crc kubenswrapper[4742]: I1004 03:11:01.956315 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:01Z","lastTransitionTime":"2025-10-04T03:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.045255 4742 generic.go:334] "Generic (PLEG): container finished" podID="87c84d43-0375-42ca-b3b3-9cadda96fcbd" containerID="7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905" exitCode=0 Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.045366 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" event={"ID":"87c84d43-0375-42ca-b3b3-9cadda96fcbd","Type":"ContainerDied","Data":"7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905"} Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.047493 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-cwm6r" event={"ID":"09a5aee9-b917-473a-9196-5cde6553c72a","Type":"ContainerStarted","Data":"a6cefd4be7726c02ec9a7b2db28ce61c5dd11b708a131e05e96595b51719209b"} Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.047608 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-cwm6r" event={"ID":"09a5aee9-b917-473a-9196-5cde6553c72a","Type":"ContainerStarted","Data":"9688d82ad1672c2fd0714b838dc0a5d5a7a8381e9af453734a0d0894fb43ef78"} Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.058365 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.058419 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.058434 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.058458 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.058474 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:02Z","lastTransitionTime":"2025-10-04T03:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.071848 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.084906 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11b178c3523269e8b4c09a1ad6f86eaee1d8f19f5056840dfa24466248889a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.099438 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.118961 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.132633 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.147828 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.161576 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.161617 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.161633 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.161657 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.161675 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:02Z","lastTransitionTime":"2025-10-04T03:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.161952 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc8m7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46de52ee-226c-4713-bd7b-961f8f43f635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pj2t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc8m7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.176543 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.192536 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.207065 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.230246 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qrtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.255689 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://261e8ace3d214aec7b737ea538ac0996cc11a5dae65eacc33dd3a3ea65ed3558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.264297 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.264345 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.264358 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.264380 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.264393 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:02Z","lastTransitionTime":"2025-10-04T03:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.281954 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.296595 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cwm6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a5aee9-b917-473a-9196-5cde6553c72a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrdmr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cwm6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.315735 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.338480 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc8m7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46de52ee-226c-4713-bd7b-961f8f43f635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pj2t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc8m7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.356211 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.367417 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.367477 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.367486 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.367502 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.367511 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:02Z","lastTransitionTime":"2025-10-04T03:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.374083 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.390481 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.405165 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.405304 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:02 crc kubenswrapper[4742]: E1004 03:11:02.405360 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:11:10.40534359 +0000 UTC m=+36.337680766 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:11:02 crc kubenswrapper[4742]: E1004 03:11:02.405407 4742 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:11:02 crc kubenswrapper[4742]: E1004 03:11:02.405493 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:11:10.405458863 +0000 UTC m=+36.337796049 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.405528 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.450250 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qrtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.470219 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.470295 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.470307 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.470329 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.470341 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:02Z","lastTransitionTime":"2025-10-04T03:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.483184 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.506465 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.506521 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.506563 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:02 crc kubenswrapper[4742]: E1004 03:11:02.506679 4742 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:11:02 crc kubenswrapper[4742]: E1004 03:11:02.506728 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:11:10.506713633 +0000 UTC m=+36.439050809 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:11:02 crc kubenswrapper[4742]: E1004 03:11:02.506728 4742 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:11:02 crc kubenswrapper[4742]: E1004 03:11:02.506776 4742 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:11:02 crc kubenswrapper[4742]: E1004 03:11:02.506789 4742 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:11:02 crc kubenswrapper[4742]: E1004 03:11:02.506851 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 03:11:10.506833146 +0000 UTC m=+36.439170372 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:11:02 crc kubenswrapper[4742]: E1004 03:11:02.506912 4742 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:11:02 crc kubenswrapper[4742]: E1004 03:11:02.506921 4742 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:11:02 crc kubenswrapper[4742]: E1004 03:11:02.506928 4742 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:11:02 crc kubenswrapper[4742]: E1004 03:11:02.506951 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 03:11:10.506941559 +0000 UTC m=+36.439278735 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.522246 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.561461 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cwm6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a5aee9-b917-473a-9196-5cde6553c72a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6cefd4be7726c02ec9a7b2db28ce61c5dd11b708a131e05e96595b51719209b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrdmr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cwm6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.573585 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.573628 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.573638 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.573657 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.573668 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:02Z","lastTransitionTime":"2025-10-04T03:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.602674 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://261e8ace3d214aec7b737ea538ac0996cc11a5dae65eacc33dd3a3ea65ed3558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.649201 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.676448 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.676499 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.676513 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.676538 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.676555 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:02Z","lastTransitionTime":"2025-10-04T03:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.688600 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.722678 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11b178c3523269e8b4c09a1ad6f86eaee1d8f19f5056840dfa24466248889a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.779318 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.779375 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.779387 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.779410 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.779424 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:02Z","lastTransitionTime":"2025-10-04T03:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.857451 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.857534 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:11:02 crc kubenswrapper[4742]: E1004 03:11:02.857620 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:11:02 crc kubenswrapper[4742]: E1004 03:11:02.857737 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.857880 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:11:02 crc kubenswrapper[4742]: E1004 03:11:02.858201 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.883239 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.883313 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.883325 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.883341 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.883354 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:02Z","lastTransitionTime":"2025-10-04T03:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.987033 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.987115 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.987132 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.987157 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:02 crc kubenswrapper[4742]: I1004 03:11:02.987171 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:02Z","lastTransitionTime":"2025-10-04T03:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.059540 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" event={"ID":"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a","Type":"ContainerStarted","Data":"e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4"} Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.063569 4742 generic.go:334] "Generic (PLEG): container finished" podID="87c84d43-0375-42ca-b3b3-9cadda96fcbd" containerID="c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7" exitCode=0 Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.063612 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" event={"ID":"87c84d43-0375-42ca-b3b3-9cadda96fcbd","Type":"ContainerDied","Data":"c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7"} Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.080063 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc8m7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46de52ee-226c-4713-bd7b-961f8f43f635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pj2t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc8m7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.090526 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.090581 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.090592 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.090612 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.090624 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:03Z","lastTransitionTime":"2025-10-04T03:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.099760 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.113993 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.130948 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.148078 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.169377 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qrtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.184406 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.192814 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.192864 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.192879 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.192902 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.192914 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:03Z","lastTransitionTime":"2025-10-04T03:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.199745 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.216471 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.229119 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://261e8ace3d214aec7b737ea538ac0996cc11a5dae65eacc33dd3a3ea65ed3558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.250011 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.263454 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cwm6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a5aee9-b917-473a-9196-5cde6553c72a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6cefd4be7726c02ec9a7b2db28ce61c5dd11b708a131e05e96595b51719209b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrdmr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cwm6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.281307 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.294812 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11b178c3523269e8b4c09a1ad6f86eaee1d8f19f5056840dfa24466248889a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.296352 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.296393 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.296406 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.296431 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.296442 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:03Z","lastTransitionTime":"2025-10-04T03:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.400569 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.400623 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.400638 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.400659 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.400674 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:03Z","lastTransitionTime":"2025-10-04T03:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.503980 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.504072 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.504093 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.504127 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.504154 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:03Z","lastTransitionTime":"2025-10-04T03:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.607412 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.607467 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.607481 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.607503 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.607518 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:03Z","lastTransitionTime":"2025-10-04T03:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.710002 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.710068 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.710081 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.710098 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.710435 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:03Z","lastTransitionTime":"2025-10-04T03:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.813221 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.813291 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.813306 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.813329 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.813343 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:03Z","lastTransitionTime":"2025-10-04T03:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.917702 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.917772 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.917791 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.917820 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:03 crc kubenswrapper[4742]: I1004 03:11:03.917840 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:03Z","lastTransitionTime":"2025-10-04T03:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.021266 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.021408 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.021438 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.021478 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.021504 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:04Z","lastTransitionTime":"2025-10-04T03:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.071107 4742 generic.go:334] "Generic (PLEG): container finished" podID="87c84d43-0375-42ca-b3b3-9cadda96fcbd" containerID="919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed" exitCode=0 Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.071160 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" event={"ID":"87c84d43-0375-42ca-b3b3-9cadda96fcbd","Type":"ContainerDied","Data":"919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed"} Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.090675 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.111643 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.128568 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.128654 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.128677 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.128721 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.128743 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:04Z","lastTransitionTime":"2025-10-04T03:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.133883 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.151570 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc8m7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46de52ee-226c-4713-bd7b-961f8f43f635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pj2t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc8m7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.168205 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.186399 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.205808 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.219432 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.233042 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.233101 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.233115 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.233173 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.233189 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:04Z","lastTransitionTime":"2025-10-04T03:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.240363 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qrtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.254634 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://261e8ace3d214aec7b737ea538ac0996cc11a5dae65eacc33dd3a3ea65ed3558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.271591 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.283229 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cwm6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a5aee9-b917-473a-9196-5cde6553c72a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6cefd4be7726c02ec9a7b2db28ce61c5dd11b708a131e05e96595b51719209b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrdmr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cwm6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.298724 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.312585 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11b178c3523269e8b4c09a1ad6f86eaee1d8f19f5056840dfa24466248889a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.336394 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.336442 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.336456 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.336479 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.336494 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:04Z","lastTransitionTime":"2025-10-04T03:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.439365 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.439417 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.439430 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.439454 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.439467 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:04Z","lastTransitionTime":"2025-10-04T03:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.542686 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.542763 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.542784 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.542813 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.542835 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:04Z","lastTransitionTime":"2025-10-04T03:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.645410 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.645460 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.645471 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.645523 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.645537 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:04Z","lastTransitionTime":"2025-10-04T03:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.750057 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.750115 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.750127 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.750151 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.750166 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:04Z","lastTransitionTime":"2025-10-04T03:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.853308 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.853401 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.853423 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.853462 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.853486 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:04Z","lastTransitionTime":"2025-10-04T03:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.857757 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.857824 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.857896 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:11:04 crc kubenswrapper[4742]: E1004 03:11:04.858051 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:11:04 crc kubenswrapper[4742]: E1004 03:11:04.858424 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:11:04 crc kubenswrapper[4742]: E1004 03:11:04.858594 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.878015 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc8m7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46de52ee-226c-4713-bd7b-961f8f43f635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pj2t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc8m7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.900524 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.923336 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.945010 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.956367 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.956415 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.956431 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.956452 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.956467 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:04Z","lastTransitionTime":"2025-10-04T03:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:04 crc kubenswrapper[4742]: I1004 03:11:04.971127 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.001196 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qrtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.020868 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.039930 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.054827 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.059474 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.059534 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.059548 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.059573 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.059589 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:05Z","lastTransitionTime":"2025-10-04T03:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.068729 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://261e8ace3d214aec7b737ea538ac0996cc11a5dae65eacc33dd3a3ea65ed3558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.080420 4742 generic.go:334] "Generic (PLEG): container finished" podID="87c84d43-0375-42ca-b3b3-9cadda96fcbd" containerID="94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a" exitCode=0 Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.080481 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" event={"ID":"87c84d43-0375-42ca-b3b3-9cadda96fcbd","Type":"ContainerDied","Data":"94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a"} Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.086413 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.090028 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.105595 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cwm6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a5aee9-b917-473a-9196-5cde6553c72a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6cefd4be7726c02ec9a7b2db28ce61c5dd11b708a131e05e96595b51719209b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrdmr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cwm6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.125655 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.141515 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11b178c3523269e8b4c09a1ad6f86eaee1d8f19f5056840dfa24466248889a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.159094 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://261e8ace3d214aec7b737ea538ac0996cc11a5dae65eacc33dd3a3ea65ed3558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.162160 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.162245 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.162260 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.162314 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.162330 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:05Z","lastTransitionTime":"2025-10-04T03:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.180260 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.194651 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cwm6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a5aee9-b917-473a-9196-5cde6553c72a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6cefd4be7726c02ec9a7b2db28ce61c5dd11b708a131e05e96595b51719209b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrdmr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cwm6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.216784 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.230613 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11b178c3523269e8b4c09a1ad6f86eaee1d8f19f5056840dfa24466248889a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.244517 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.260932 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.265109 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.265167 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.265179 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.265197 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.265208 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:05Z","lastTransitionTime":"2025-10-04T03:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.275723 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.289935 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.303658 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc8m7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46de52ee-226c-4713-bd7b-961f8f43f635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pj2t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc8m7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.317922 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.330761 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.346362 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.366768 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qrtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.367800 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.367837 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.367846 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.367867 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.367880 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:05Z","lastTransitionTime":"2025-10-04T03:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.470814 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.470860 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.470871 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.470887 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.470899 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:05Z","lastTransitionTime":"2025-10-04T03:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.574432 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.574483 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.574494 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.574512 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.574524 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:05Z","lastTransitionTime":"2025-10-04T03:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.678075 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.678113 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.678122 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.678138 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.678148 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:05Z","lastTransitionTime":"2025-10-04T03:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.784129 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.784191 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.784203 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.784223 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.784238 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:05Z","lastTransitionTime":"2025-10-04T03:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.887484 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.887546 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.887562 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.887582 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.887599 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:05Z","lastTransitionTime":"2025-10-04T03:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.990391 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.990809 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.990828 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.990849 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:05 crc kubenswrapper[4742]: I1004 03:11:05.990863 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:05Z","lastTransitionTime":"2025-10-04T03:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.089472 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" event={"ID":"87c84d43-0375-42ca-b3b3-9cadda96fcbd","Type":"ContainerStarted","Data":"dd1d6092975dd4703ff4930e2eafaf464f221510fddeda0f977a73c548c305eb"} Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.093246 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.093321 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.093336 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.093355 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.093369 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:06Z","lastTransitionTime":"2025-10-04T03:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.097050 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" event={"ID":"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a","Type":"ContainerStarted","Data":"f556ea9ccf10b49a8b1097267b402b532f8df5b1f47d7bcd0a45eb56eeaaedfc"} Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.097629 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.097692 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.110417 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd1d6092975dd4703ff4930e2eafaf464f221510fddeda0f977a73c548c305eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.126756 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cwm6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a5aee9-b917-473a-9196-5cde6553c72a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6cefd4be7726c02ec9a7b2db28ce61c5dd11b708a131e05e96595b51719209b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrdmr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cwm6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.168002 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.168116 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.179600 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://261e8ace3d214aec7b737ea538ac0996cc11a5dae65eacc33dd3a3ea65ed3558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.196651 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.196734 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.196753 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.196783 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.196799 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:06Z","lastTransitionTime":"2025-10-04T03:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.196897 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11b178c3523269e8b4c09a1ad6f86eaee1d8f19f5056840dfa24466248889a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.212824 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.226915 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.242564 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.258201 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc8m7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46de52ee-226c-4713-bd7b-961f8f43f635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pj2t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc8m7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.275142 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.290173 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.300461 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.300529 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.300542 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.300564 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.300577 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:06Z","lastTransitionTime":"2025-10-04T03:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.302904 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.317457 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.345698 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qrtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.359717 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.373950 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc8m7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46de52ee-226c-4713-bd7b-961f8f43f635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pj2t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc8m7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.389414 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.403043 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.403102 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.403118 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.403139 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.403151 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:06Z","lastTransitionTime":"2025-10-04T03:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.405923 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.420179 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.437411 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.463321 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f556ea9ccf10b49a8b1097267b402b532f8df5b1f47d7bcd0a45eb56eeaaedfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qrtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.478772 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.492740 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.505839 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.505889 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.505901 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.505919 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.505933 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:06Z","lastTransitionTime":"2025-10-04T03:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.509128 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.526740 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://261e8ace3d214aec7b737ea538ac0996cc11a5dae65eacc33dd3a3ea65ed3558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.545563 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd1d6092975dd4703ff4930e2eafaf464f221510fddeda0f977a73c548c305eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.559469 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cwm6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a5aee9-b917-473a-9196-5cde6553c72a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6cefd4be7726c02ec9a7b2db28ce61c5dd11b708a131e05e96595b51719209b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrdmr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cwm6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.574655 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.587884 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11b178c3523269e8b4c09a1ad6f86eaee1d8f19f5056840dfa24466248889a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.609649 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.609698 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.609716 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.609736 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.609750 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:06Z","lastTransitionTime":"2025-10-04T03:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.713448 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.713530 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.713554 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.713586 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.713610 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:06Z","lastTransitionTime":"2025-10-04T03:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.816360 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.816420 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.816430 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.816449 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.816464 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:06Z","lastTransitionTime":"2025-10-04T03:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.858014 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.858064 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.858018 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:11:06 crc kubenswrapper[4742]: E1004 03:11:06.858194 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:11:06 crc kubenswrapper[4742]: E1004 03:11:06.858394 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:11:06 crc kubenswrapper[4742]: E1004 03:11:06.858621 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.918736 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.918792 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.918804 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.918822 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:06 crc kubenswrapper[4742]: I1004 03:11:06.918834 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:06Z","lastTransitionTime":"2025-10-04T03:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.021354 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.021386 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.021396 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.021410 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.021420 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:07Z","lastTransitionTime":"2025-10-04T03:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.100825 4742 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.124185 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.124246 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.124265 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.124323 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.124342 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:07Z","lastTransitionTime":"2025-10-04T03:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.226937 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.227014 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.227040 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.227068 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.227092 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:07Z","lastTransitionTime":"2025-10-04T03:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.331388 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.331442 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.331455 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.331476 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.331490 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:07Z","lastTransitionTime":"2025-10-04T03:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.434626 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.434679 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.434703 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.434727 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.434756 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:07Z","lastTransitionTime":"2025-10-04T03:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.538063 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.538146 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.538173 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.538205 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.538231 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:07Z","lastTransitionTime":"2025-10-04T03:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.641491 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.641548 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.641561 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.641580 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.641594 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:07Z","lastTransitionTime":"2025-10-04T03:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.744757 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.744800 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.744810 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.744827 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.744837 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:07Z","lastTransitionTime":"2025-10-04T03:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.848457 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.848537 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.848557 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.848586 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.848608 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:07Z","lastTransitionTime":"2025-10-04T03:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.952623 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.952731 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.952765 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.952807 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:07 crc kubenswrapper[4742]: I1004 03:11:07.952833 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:07Z","lastTransitionTime":"2025-10-04T03:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.061684 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.061753 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.061771 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.061795 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.062011 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:08Z","lastTransitionTime":"2025-10-04T03:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.103448 4742 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.165252 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.165352 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.165371 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.165397 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.165414 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:08Z","lastTransitionTime":"2025-10-04T03:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.269347 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.269416 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.269431 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.269459 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.269475 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:08Z","lastTransitionTime":"2025-10-04T03:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.372690 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.372777 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.372795 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.372819 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.372836 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:08Z","lastTransitionTime":"2025-10-04T03:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.475833 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.475895 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.475913 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.475969 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.475985 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:08Z","lastTransitionTime":"2025-10-04T03:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.579713 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.579765 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.579788 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.579816 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.579827 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:08Z","lastTransitionTime":"2025-10-04T03:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.683438 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.683497 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.683509 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.683530 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.683544 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:08Z","lastTransitionTime":"2025-10-04T03:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.787055 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.787125 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.787140 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.787162 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.787176 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:08Z","lastTransitionTime":"2025-10-04T03:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.858020 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:11:08 crc kubenswrapper[4742]: E1004 03:11:08.858397 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.859535 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.859620 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:11:08 crc kubenswrapper[4742]: E1004 03:11:08.859722 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:11:08 crc kubenswrapper[4742]: E1004 03:11:08.859874 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.890600 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.890658 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.890671 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.890692 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.890707 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:08Z","lastTransitionTime":"2025-10-04T03:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.993826 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.993888 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.993913 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.993945 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:08 crc kubenswrapper[4742]: I1004 03:11:08.993961 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:08Z","lastTransitionTime":"2025-10-04T03:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.096197 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.096291 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.096354 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.096387 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.096406 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:09Z","lastTransitionTime":"2025-10-04T03:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.108523 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qrtl_0cfc6e64-ba08-4c44-a18e-3b93ab792a6a/ovnkube-controller/0.log" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.113337 4742 generic.go:334] "Generic (PLEG): container finished" podID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerID="f556ea9ccf10b49a8b1097267b402b532f8df5b1f47d7bcd0a45eb56eeaaedfc" exitCode=1 Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.113375 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" event={"ID":"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a","Type":"ContainerDied","Data":"f556ea9ccf10b49a8b1097267b402b532f8df5b1f47d7bcd0a45eb56eeaaedfc"} Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.114082 4742 scope.go:117] "RemoveContainer" containerID="f556ea9ccf10b49a8b1097267b402b532f8df5b1f47d7bcd0a45eb56eeaaedfc" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.135360 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://261e8ace3d214aec7b737ea538ac0996cc11a5dae65eacc33dd3a3ea65ed3558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:09Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.158645 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd1d6092975dd4703ff4930e2eafaf464f221510fddeda0f977a73c548c305eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:09Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.178958 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cwm6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a5aee9-b917-473a-9196-5cde6553c72a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6cefd4be7726c02ec9a7b2db28ce61c5dd11b708a131e05e96595b51719209b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrdmr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cwm6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:09Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.199762 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.199809 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.199826 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.199852 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.199873 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:09Z","lastTransitionTime":"2025-10-04T03:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.202754 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:09Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.218086 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11b178c3523269e8b4c09a1ad6f86eaee1d8f19f5056840dfa24466248889a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:09Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.240672 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:09Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.261567 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:09Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.277826 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:09Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.298632 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:09Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.302404 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.302432 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.302443 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.302460 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.302473 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:09Z","lastTransitionTime":"2025-10-04T03:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.315716 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc8m7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46de52ee-226c-4713-bd7b-961f8f43f635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pj2t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc8m7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:09Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.331406 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:09Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.345243 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:09Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.360450 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:09Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.383740 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f556ea9ccf10b49a8b1097267b402b532f8df5b1f47d7bcd0a45eb56eeaaedfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f556ea9ccf10b49a8b1097267b402b532f8df5b1f47d7bcd0a45eb56eeaaedfc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:11:08Z\\\",\\\"message\\\":\\\"8.183889 6051 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:11:08.184000 6051 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:11:08.184057 6051 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:11:08.184239 6051 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:11:08.184529 6051 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:11:08.184927 6051 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:11:08.185386 6051 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 03:11:08.185496 6051 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:11:08.185515 6051 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 03:11:08.185549 6051 factory.go:656] Stopping watch factory\\\\nI1004 03:11:08.185554 6051 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 03:11:08.185557 6051 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:11:08.185600 6051 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qrtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:09Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.404876 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.404955 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.404975 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.405006 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.405033 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:09Z","lastTransitionTime":"2025-10-04T03:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.507910 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.507941 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.507950 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.507961 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.507971 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:09Z","lastTransitionTime":"2025-10-04T03:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.609964 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.610022 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.610035 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.610057 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.610070 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:09Z","lastTransitionTime":"2025-10-04T03:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.712853 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.712913 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.712932 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.712956 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.712976 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:09Z","lastTransitionTime":"2025-10-04T03:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.815833 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.815891 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.815905 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.815927 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.815944 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:09Z","lastTransitionTime":"2025-10-04T03:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.918398 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.918479 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.918500 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.918529 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:09 crc kubenswrapper[4742]: I1004 03:11:09.918548 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:09Z","lastTransitionTime":"2025-10-04T03:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.020826 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.020863 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.020873 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.020887 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.020898 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:10Z","lastTransitionTime":"2025-10-04T03:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.118742 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qrtl_0cfc6e64-ba08-4c44-a18e-3b93ab792a6a/ovnkube-controller/1.log" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.119509 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qrtl_0cfc6e64-ba08-4c44-a18e-3b93ab792a6a/ovnkube-controller/0.log" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.122372 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.122394 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.122402 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.122417 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.122426 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:10Z","lastTransitionTime":"2025-10-04T03:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.123668 4742 generic.go:334] "Generic (PLEG): container finished" podID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerID="53d71d0eb0f06d88c47adfbd6bcb96a5c47a273919690b3cae9be51fff8b5a67" exitCode=1 Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.123699 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" event={"ID":"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a","Type":"ContainerDied","Data":"53d71d0eb0f06d88c47adfbd6bcb96a5c47a273919690b3cae9be51fff8b5a67"} Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.123735 4742 scope.go:117] "RemoveContainer" containerID="f556ea9ccf10b49a8b1097267b402b532f8df5b1f47d7bcd0a45eb56eeaaedfc" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.125211 4742 scope.go:117] "RemoveContainer" containerID="53d71d0eb0f06d88c47adfbd6bcb96a5c47a273919690b3cae9be51fff8b5a67" Oct 04 03:11:10 crc kubenswrapper[4742]: E1004 03:11:10.125569 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-9qrtl_openshift-ovn-kubernetes(0cfc6e64-ba08-4c44-a18e-3b93ab792a6a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.144044 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://261e8ace3d214aec7b737ea538ac0996cc11a5dae65eacc33dd3a3ea65ed3558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:10Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.170867 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd1d6092975dd4703ff4930e2eafaf464f221510fddeda0f977a73c548c305eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:10Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.185865 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cwm6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a5aee9-b917-473a-9196-5cde6553c72a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6cefd4be7726c02ec9a7b2db28ce61c5dd11b708a131e05e96595b51719209b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrdmr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cwm6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:10Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.207881 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:10Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.224347 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11b178c3523269e8b4c09a1ad6f86eaee1d8f19f5056840dfa24466248889a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:10Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.224877 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.224925 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.224942 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.224961 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.224974 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:10Z","lastTransitionTime":"2025-10-04T03:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.242144 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:10Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.258425 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:10Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.279061 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:10Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.296967 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:10Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.313434 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc8m7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46de52ee-226c-4713-bd7b-961f8f43f635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pj2t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc8m7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:10Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.326515 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:10Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.327958 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.328009 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.328022 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.328044 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.328055 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:10Z","lastTransitionTime":"2025-10-04T03:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.342690 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:10Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.362319 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:10Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.386516 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53d71d0eb0f06d88c47adfbd6bcb96a5c47a273919690b3cae9be51fff8b5a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f556ea9ccf10b49a8b1097267b402b532f8df5b1f47d7bcd0a45eb56eeaaedfc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:11:08Z\\\",\\\"message\\\":\\\"8.183889 6051 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:11:08.184000 6051 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:11:08.184057 6051 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:11:08.184239 6051 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:11:08.184529 6051 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:11:08.184927 6051 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:11:08.185386 6051 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 03:11:08.185496 6051 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:11:08.185515 6051 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 03:11:08.185549 6051 factory.go:656] Stopping watch factory\\\\nI1004 03:11:08.185554 6051 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 03:11:08.185557 6051 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:11:08.185600 6051 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d71d0eb0f06d88c47adfbd6bcb96a5c47a273919690b3cae9be51fff8b5a67\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:11:09Z\\\",\\\"message\\\":\\\"nd\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/console\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console/console_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/console\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.194\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1004 03:11:09.936938 6184 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Inte\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qrtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:10Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.430978 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.431027 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.431036 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.431051 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.431061 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:10Z","lastTransitionTime":"2025-10-04T03:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.498787 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.498937 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:10 crc kubenswrapper[4742]: E1004 03:11:10.499042 4742 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:11:10 crc kubenswrapper[4742]: E1004 03:11:10.499123 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:11:26.499053379 +0000 UTC m=+52.431390605 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:11:10 crc kubenswrapper[4742]: E1004 03:11:10.499216 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:11:26.499193592 +0000 UTC m=+52.431531008 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.534551 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.534596 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.534615 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.534634 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.534647 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:10Z","lastTransitionTime":"2025-10-04T03:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.597485 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.597533 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.597543 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.597560 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.597620 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:10Z","lastTransitionTime":"2025-10-04T03:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.599225 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.599262 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.599291 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:11:10 crc kubenswrapper[4742]: E1004 03:11:10.599400 4742 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:11:10 crc kubenswrapper[4742]: E1004 03:11:10.599418 4742 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:11:10 crc kubenswrapper[4742]: E1004 03:11:10.599427 4742 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:11:10 crc kubenswrapper[4742]: E1004 03:11:10.599398 4742 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:11:10 crc kubenswrapper[4742]: E1004 03:11:10.599511 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 03:11:26.599497008 +0000 UTC m=+52.531834184 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:11:10 crc kubenswrapper[4742]: E1004 03:11:10.599549 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:11:26.599534149 +0000 UTC m=+52.531871325 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:11:10 crc kubenswrapper[4742]: E1004 03:11:10.599688 4742 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:11:10 crc kubenswrapper[4742]: E1004 03:11:10.599753 4742 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:11:10 crc kubenswrapper[4742]: E1004 03:11:10.599788 4742 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:11:10 crc kubenswrapper[4742]: E1004 03:11:10.599933 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 03:11:26.599890218 +0000 UTC m=+52.532227444 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:11:10 crc kubenswrapper[4742]: E1004 03:11:10.620094 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:10Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.625113 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.625182 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.625204 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.625238 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.625263 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:10Z","lastTransitionTime":"2025-10-04T03:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:10 crc kubenswrapper[4742]: E1004 03:11:10.643104 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:10Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.648319 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.648395 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.648416 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.648447 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.648468 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:10Z","lastTransitionTime":"2025-10-04T03:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:10 crc kubenswrapper[4742]: E1004 03:11:10.668417 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:10Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.673570 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.673621 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.673668 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.673689 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.673701 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:10Z","lastTransitionTime":"2025-10-04T03:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:10 crc kubenswrapper[4742]: E1004 03:11:10.687855 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:10Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.693087 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.693148 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.693167 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.693193 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.693211 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:10Z","lastTransitionTime":"2025-10-04T03:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:10 crc kubenswrapper[4742]: E1004 03:11:10.708741 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:10Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:10 crc kubenswrapper[4742]: E1004 03:11:10.708904 4742 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.710910 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.710970 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.710983 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.711002 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.711019 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:10Z","lastTransitionTime":"2025-10-04T03:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.814111 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.814176 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.814189 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.814212 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.814226 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:10Z","lastTransitionTime":"2025-10-04T03:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.857414 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.857470 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:10 crc kubenswrapper[4742]: E1004 03:11:10.857578 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.857668 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:11:10 crc kubenswrapper[4742]: E1004 03:11:10.857896 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:11:10 crc kubenswrapper[4742]: E1004 03:11:10.858035 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.918184 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.918241 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.918254 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.918300 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:10 crc kubenswrapper[4742]: I1004 03:11:10.918316 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:10Z","lastTransitionTime":"2025-10-04T03:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.021164 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.021222 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.021234 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.021256 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.021274 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:11Z","lastTransitionTime":"2025-10-04T03:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.124741 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.124787 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.124796 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.124809 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.124819 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:11Z","lastTransitionTime":"2025-10-04T03:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.128988 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qrtl_0cfc6e64-ba08-4c44-a18e-3b93ab792a6a/ovnkube-controller/1.log" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.227266 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.227365 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.227384 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.227414 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.227434 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:11Z","lastTransitionTime":"2025-10-04T03:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.329785 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.329852 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.329864 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.329882 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.329893 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:11Z","lastTransitionTime":"2025-10-04T03:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.432301 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.432352 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.432365 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.432383 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.432395 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:11Z","lastTransitionTime":"2025-10-04T03:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.460923 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4"] Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.461465 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.464061 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.465432 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.482771 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:11Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.494032 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11b178c3523269e8b4c09a1ad6f86eaee1d8f19f5056840dfa24466248889a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:11Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.508588 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:11Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.521975 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:11Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.535144 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.535190 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.535200 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.535218 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.535229 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:11Z","lastTransitionTime":"2025-10-04T03:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.536558 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:11Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.555309 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc8m7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46de52ee-226c-4713-bd7b-961f8f43f635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pj2t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc8m7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:11Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.569434 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2694d67-7fb5-446f-a590-6113e9131921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwzp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:11Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.583758 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:11Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.602527 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:11Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.609474 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b2694d67-7fb5-446f-a590-6113e9131921-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-pwzp4\" (UID: \"b2694d67-7fb5-446f-a590-6113e9131921\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.609568 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b2694d67-7fb5-446f-a590-6113e9131921-env-overrides\") pod \"ovnkube-control-plane-749d76644c-pwzp4\" (UID: \"b2694d67-7fb5-446f-a590-6113e9131921\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.609612 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b2694d67-7fb5-446f-a590-6113e9131921-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-pwzp4\" (UID: \"b2694d67-7fb5-446f-a590-6113e9131921\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.609649 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgssc\" (UniqueName: \"kubernetes.io/projected/b2694d67-7fb5-446f-a590-6113e9131921-kube-api-access-mgssc\") pod \"ovnkube-control-plane-749d76644c-pwzp4\" (UID: \"b2694d67-7fb5-446f-a590-6113e9131921\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.618014 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:11Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.633717 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:11Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.638596 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.638648 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.638660 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.638682 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.638692 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:11Z","lastTransitionTime":"2025-10-04T03:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.671677 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53d71d0eb0f06d88c47adfbd6bcb96a5c47a273919690b3cae9be51fff8b5a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f556ea9ccf10b49a8b1097267b402b532f8df5b1f47d7bcd0a45eb56eeaaedfc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:11:08Z\\\",\\\"message\\\":\\\"8.183889 6051 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:11:08.184000 6051 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:11:08.184057 6051 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:11:08.184239 6051 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:11:08.184529 6051 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:11:08.184927 6051 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:11:08.185386 6051 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 03:11:08.185496 6051 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:11:08.185515 6051 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 03:11:08.185549 6051 factory.go:656] Stopping watch factory\\\\nI1004 03:11:08.185554 6051 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 03:11:08.185557 6051 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:11:08.185600 6051 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d71d0eb0f06d88c47adfbd6bcb96a5c47a273919690b3cae9be51fff8b5a67\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:11:09Z\\\",\\\"message\\\":\\\"nd\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/console\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console/console_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/console\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.194\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1004 03:11:09.936938 6184 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Inte\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qrtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:11Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.687790 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://261e8ace3d214aec7b737ea538ac0996cc11a5dae65eacc33dd3a3ea65ed3558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:11Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.710425 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b2694d67-7fb5-446f-a590-6113e9131921-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-pwzp4\" (UID: \"b2694d67-7fb5-446f-a590-6113e9131921\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.710401 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd1d6092975dd4703ff4930e2eafaf464f221510fddeda0f977a73c548c305eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:11Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.710499 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b2694d67-7fb5-446f-a590-6113e9131921-env-overrides\") pod \"ovnkube-control-plane-749d76644c-pwzp4\" (UID: \"b2694d67-7fb5-446f-a590-6113e9131921\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.710795 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b2694d67-7fb5-446f-a590-6113e9131921-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-pwzp4\" (UID: \"b2694d67-7fb5-446f-a590-6113e9131921\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.710825 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgssc\" (UniqueName: \"kubernetes.io/projected/b2694d67-7fb5-446f-a590-6113e9131921-kube-api-access-mgssc\") pod \"ovnkube-control-plane-749d76644c-pwzp4\" (UID: \"b2694d67-7fb5-446f-a590-6113e9131921\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.711329 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b2694d67-7fb5-446f-a590-6113e9131921-env-overrides\") pod \"ovnkube-control-plane-749d76644c-pwzp4\" (UID: \"b2694d67-7fb5-446f-a590-6113e9131921\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.712322 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b2694d67-7fb5-446f-a590-6113e9131921-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-pwzp4\" (UID: \"b2694d67-7fb5-446f-a590-6113e9131921\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.722491 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b2694d67-7fb5-446f-a590-6113e9131921-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-pwzp4\" (UID: \"b2694d67-7fb5-446f-a590-6113e9131921\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.727850 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cwm6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a5aee9-b917-473a-9196-5cde6553c72a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6cefd4be7726c02ec9a7b2db28ce61c5dd11b708a131e05e96595b51719209b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrdmr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cwm6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:11Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.741051 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgssc\" (UniqueName: \"kubernetes.io/projected/b2694d67-7fb5-446f-a590-6113e9131921-kube-api-access-mgssc\") pod \"ovnkube-control-plane-749d76644c-pwzp4\" (UID: \"b2694d67-7fb5-446f-a590-6113e9131921\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.741212 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.741232 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.741242 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.741257 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.741267 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:11Z","lastTransitionTime":"2025-10-04T03:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.775786 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.844488 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.844553 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.844570 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.844598 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.844618 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:11Z","lastTransitionTime":"2025-10-04T03:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.947516 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.947560 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.947571 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.947591 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:11 crc kubenswrapper[4742]: I1004 03:11:11.947602 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:11Z","lastTransitionTime":"2025-10-04T03:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.051558 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.051623 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.051641 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.051672 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.051694 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:12Z","lastTransitionTime":"2025-10-04T03:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.138508 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" event={"ID":"b2694d67-7fb5-446f-a590-6113e9131921","Type":"ContainerStarted","Data":"c1070bde4d35430f4c712d63ede34a428ec20a277ce0409dd3e69dead27c3f98"} Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.138576 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" event={"ID":"b2694d67-7fb5-446f-a590-6113e9131921","Type":"ContainerStarted","Data":"f4b29154af28bbb704eeb5fde7a5c44db624f1cb35d79942c62d5ca9a03e697b"} Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.138592 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" event={"ID":"b2694d67-7fb5-446f-a590-6113e9131921","Type":"ContainerStarted","Data":"8fba6760461d8ed9027fbe4ad87d559051bbb19527bd1c2a8bbb9c2525a8b948"} Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.152951 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://261e8ace3d214aec7b737ea538ac0996cc11a5dae65eacc33dd3a3ea65ed3558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:12Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.155351 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.155417 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.155504 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.155531 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.155573 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:12Z","lastTransitionTime":"2025-10-04T03:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.170447 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd1d6092975dd4703ff4930e2eafaf464f221510fddeda0f977a73c548c305eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:12Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.187157 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cwm6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a5aee9-b917-473a-9196-5cde6553c72a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6cefd4be7726c02ec9a7b2db28ce61c5dd11b708a131e05e96595b51719209b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrdmr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cwm6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:12Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.208797 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:12Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.223264 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11b178c3523269e8b4c09a1ad6f86eaee1d8f19f5056840dfa24466248889a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:12Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.244553 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:12Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.258994 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.259050 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.259071 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.259099 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.259120 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:12Z","lastTransitionTime":"2025-10-04T03:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.262075 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:12Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.280621 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:12Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.302138 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:12Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.316601 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc8m7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46de52ee-226c-4713-bd7b-961f8f43f635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pj2t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc8m7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:12Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.332220 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2694d67-7fb5-446f-a590-6113e9131921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b29154af28bbb704eeb5fde7a5c44db624f1cb35d79942c62d5ca9a03e697b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1070bde4d35430f4c712d63ede34a428ec20a277ce0409dd3e69dead27c3f98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwzp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:12Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.346760 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:12Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.361483 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:12Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.362377 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.362421 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.362431 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.362450 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.362460 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:12Z","lastTransitionTime":"2025-10-04T03:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.375421 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:12Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.395231 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53d71d0eb0f06d88c47adfbd6bcb96a5c47a273919690b3cae9be51fff8b5a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f556ea9ccf10b49a8b1097267b402b532f8df5b1f47d7bcd0a45eb56eeaaedfc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:11:08Z\\\",\\\"message\\\":\\\"8.183889 6051 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:11:08.184000 6051 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:11:08.184057 6051 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:11:08.184239 6051 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:11:08.184529 6051 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:11:08.184927 6051 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:11:08.185386 6051 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 03:11:08.185496 6051 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:11:08.185515 6051 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 03:11:08.185549 6051 factory.go:656] Stopping watch factory\\\\nI1004 03:11:08.185554 6051 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 03:11:08.185557 6051 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:11:08.185600 6051 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d71d0eb0f06d88c47adfbd6bcb96a5c47a273919690b3cae9be51fff8b5a67\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:11:09Z\\\",\\\"message\\\":\\\"nd\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/console\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console/console_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/console\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.194\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1004 03:11:09.936938 6184 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Inte\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qrtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:12Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.464818 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.464916 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.464944 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.464978 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.465005 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:12Z","lastTransitionTime":"2025-10-04T03:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.569191 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.569240 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.569251 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.569291 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.569305 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:12Z","lastTransitionTime":"2025-10-04T03:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.575331 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-cd46x"] Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.575901 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:11:12 crc kubenswrapper[4742]: E1004 03:11:12.575985 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.592105 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:12Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.605779 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11b178c3523269e8b4c09a1ad6f86eaee1d8f19f5056840dfa24466248889a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:12Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.620295 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:12Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.638367 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:12Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.658526 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:12Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.673009 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.673067 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.673086 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.673110 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.673129 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:12Z","lastTransitionTime":"2025-10-04T03:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.675317 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:12Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.689702 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc8m7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46de52ee-226c-4713-bd7b-961f8f43f635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pj2t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc8m7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:12Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.701662 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2694d67-7fb5-446f-a590-6113e9131921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b29154af28bbb704eeb5fde7a5c44db624f1cb35d79942c62d5ca9a03e697b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1070bde4d35430f4c712d63ede34a428ec20a277ce0409dd3e69dead27c3f98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwzp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:12Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.719219 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:12Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.722885 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0302950c-deb3-420a-b798-c99459e62b3f-metrics-certs\") pod \"network-metrics-daemon-cd46x\" (UID: \"0302950c-deb3-420a-b798-c99459e62b3f\") " pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.722956 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fls9p\" (UniqueName: \"kubernetes.io/projected/0302950c-deb3-420a-b798-c99459e62b3f-kube-api-access-fls9p\") pod \"network-metrics-daemon-cd46x\" (UID: \"0302950c-deb3-420a-b798-c99459e62b3f\") " pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.736940 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:12Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.753385 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:12Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.776407 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.776504 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.776518 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.776540 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.776557 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:12Z","lastTransitionTime":"2025-10-04T03:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.779738 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53d71d0eb0f06d88c47adfbd6bcb96a5c47a273919690b3cae9be51fff8b5a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f556ea9ccf10b49a8b1097267b402b532f8df5b1f47d7bcd0a45eb56eeaaedfc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:11:08Z\\\",\\\"message\\\":\\\"8.183889 6051 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:11:08.184000 6051 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:11:08.184057 6051 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:11:08.184239 6051 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:11:08.184529 6051 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:11:08.184927 6051 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:11:08.185386 6051 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 03:11:08.185496 6051 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:11:08.185515 6051 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 03:11:08.185549 6051 factory.go:656] Stopping watch factory\\\\nI1004 03:11:08.185554 6051 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 03:11:08.185557 6051 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:11:08.185600 6051 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d71d0eb0f06d88c47adfbd6bcb96a5c47a273919690b3cae9be51fff8b5a67\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:11:09Z\\\",\\\"message\\\":\\\"nd\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/console\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console/console_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/console\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.194\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1004 03:11:09.936938 6184 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Inte\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qrtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:12Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.799876 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://261e8ace3d214aec7b737ea538ac0996cc11a5dae65eacc33dd3a3ea65ed3558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:12Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.824217 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fls9p\" (UniqueName: \"kubernetes.io/projected/0302950c-deb3-420a-b798-c99459e62b3f-kube-api-access-fls9p\") pod \"network-metrics-daemon-cd46x\" (UID: \"0302950c-deb3-420a-b798-c99459e62b3f\") " pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.824393 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0302950c-deb3-420a-b798-c99459e62b3f-metrics-certs\") pod \"network-metrics-daemon-cd46x\" (UID: \"0302950c-deb3-420a-b798-c99459e62b3f\") " pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:11:12 crc kubenswrapper[4742]: E1004 03:11:12.824634 4742 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:11:12 crc kubenswrapper[4742]: E1004 03:11:12.824771 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0302950c-deb3-420a-b798-c99459e62b3f-metrics-certs podName:0302950c-deb3-420a-b798-c99459e62b3f nodeName:}" failed. No retries permitted until 2025-10-04 03:11:13.324740708 +0000 UTC m=+39.257077884 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0302950c-deb3-420a-b798-c99459e62b3f-metrics-certs") pod "network-metrics-daemon-cd46x" (UID: "0302950c-deb3-420a-b798-c99459e62b3f") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.826576 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd1d6092975dd4703ff4930e2eafaf464f221510fddeda0f977a73c548c305eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:12Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.839404 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cwm6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a5aee9-b917-473a-9196-5cde6553c72a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6cefd4be7726c02ec9a7b2db28ce61c5dd11b708a131e05e96595b51719209b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrdmr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cwm6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:12Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.851361 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cd46x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0302950c-deb3-420a-b798-c99459e62b3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fls9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fls9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:12Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cd46x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:12Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.859474 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.859664 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.859779 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:11:12 crc kubenswrapper[4742]: E1004 03:11:12.860034 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:11:12 crc kubenswrapper[4742]: E1004 03:11:12.860123 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:11:12 crc kubenswrapper[4742]: E1004 03:11:12.859946 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.860210 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fls9p\" (UniqueName: \"kubernetes.io/projected/0302950c-deb3-420a-b798-c99459e62b3f-kube-api-access-fls9p\") pod \"network-metrics-daemon-cd46x\" (UID: \"0302950c-deb3-420a-b798-c99459e62b3f\") " pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.879329 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.879386 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.879399 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.879418 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.879430 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:12Z","lastTransitionTime":"2025-10-04T03:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.982294 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.982616 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.982719 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.982819 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:12 crc kubenswrapper[4742]: I1004 03:11:12.982912 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:12Z","lastTransitionTime":"2025-10-04T03:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.086603 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.086682 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.086703 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.086785 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.086806 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:13Z","lastTransitionTime":"2025-10-04T03:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.190222 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.190329 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.190350 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.190377 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.190395 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:13Z","lastTransitionTime":"2025-10-04T03:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.293404 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.293491 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.293509 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.293531 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.293543 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:13Z","lastTransitionTime":"2025-10-04T03:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.331551 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0302950c-deb3-420a-b798-c99459e62b3f-metrics-certs\") pod \"network-metrics-daemon-cd46x\" (UID: \"0302950c-deb3-420a-b798-c99459e62b3f\") " pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:11:13 crc kubenswrapper[4742]: E1004 03:11:13.331841 4742 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:11:13 crc kubenswrapper[4742]: E1004 03:11:13.331988 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0302950c-deb3-420a-b798-c99459e62b3f-metrics-certs podName:0302950c-deb3-420a-b798-c99459e62b3f nodeName:}" failed. No retries permitted until 2025-10-04 03:11:14.331957971 +0000 UTC m=+40.264295157 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0302950c-deb3-420a-b798-c99459e62b3f-metrics-certs") pod "network-metrics-daemon-cd46x" (UID: "0302950c-deb3-420a-b798-c99459e62b3f") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.395787 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.395833 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.395867 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.395885 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.395895 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:13Z","lastTransitionTime":"2025-10-04T03:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.498754 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.498819 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.498832 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.498850 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.498863 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:13Z","lastTransitionTime":"2025-10-04T03:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.601586 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.601643 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.601655 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.601676 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.601691 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:13Z","lastTransitionTime":"2025-10-04T03:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.705335 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.705402 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.705415 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.705440 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.705455 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:13Z","lastTransitionTime":"2025-10-04T03:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.807737 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.807830 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.807858 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.807891 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.807919 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:13Z","lastTransitionTime":"2025-10-04T03:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.858230 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:11:13 crc kubenswrapper[4742]: E1004 03:11:13.858452 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.911531 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.911579 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.911592 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.911614 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:13 crc kubenswrapper[4742]: I1004 03:11:13.911627 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:13Z","lastTransitionTime":"2025-10-04T03:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.013759 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.013808 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.013819 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.013837 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.013849 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:14Z","lastTransitionTime":"2025-10-04T03:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.117316 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.117383 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.117399 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.117422 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.117436 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:14Z","lastTransitionTime":"2025-10-04T03:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.221013 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.221071 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.221082 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.221099 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.221110 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:14Z","lastTransitionTime":"2025-10-04T03:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.323758 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.323807 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.323816 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.323831 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.323842 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:14Z","lastTransitionTime":"2025-10-04T03:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.345045 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0302950c-deb3-420a-b798-c99459e62b3f-metrics-certs\") pod \"network-metrics-daemon-cd46x\" (UID: \"0302950c-deb3-420a-b798-c99459e62b3f\") " pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:11:14 crc kubenswrapper[4742]: E1004 03:11:14.345350 4742 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:11:14 crc kubenswrapper[4742]: E1004 03:11:14.345476 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0302950c-deb3-420a-b798-c99459e62b3f-metrics-certs podName:0302950c-deb3-420a-b798-c99459e62b3f nodeName:}" failed. No retries permitted until 2025-10-04 03:11:16.345444832 +0000 UTC m=+42.277782189 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0302950c-deb3-420a-b798-c99459e62b3f-metrics-certs") pod "network-metrics-daemon-cd46x" (UID: "0302950c-deb3-420a-b798-c99459e62b3f") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.427144 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.427235 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.427256 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.427334 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.427358 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:14Z","lastTransitionTime":"2025-10-04T03:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.530817 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.530888 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.530907 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.530934 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.530951 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:14Z","lastTransitionTime":"2025-10-04T03:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.633199 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.633315 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.633341 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.633374 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.633397 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:14Z","lastTransitionTime":"2025-10-04T03:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.737016 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.737077 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.737095 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.737118 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.737130 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:14Z","lastTransitionTime":"2025-10-04T03:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.841057 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.841127 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.841143 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.841170 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.841185 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:14Z","lastTransitionTime":"2025-10-04T03:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.857982 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.857994 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.858134 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:11:14 crc kubenswrapper[4742]: E1004 03:11:14.858204 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:11:14 crc kubenswrapper[4742]: E1004 03:11:14.858300 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:11:14 crc kubenswrapper[4742]: E1004 03:11:14.858387 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.878376 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2694d67-7fb5-446f-a590-6113e9131921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b29154af28bbb704eeb5fde7a5c44db624f1cb35d79942c62d5ca9a03e697b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1070bde4d35430f4c712d63ede34a428ec20a277ce0409dd3e69dead27c3f98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwzp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:14Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.897693 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:14Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.918107 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:14Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.940161 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:14Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.944694 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.944741 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.944755 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.944782 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.944801 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:14Z","lastTransitionTime":"2025-10-04T03:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.965420 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:14Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:14 crc kubenswrapper[4742]: I1004 03:11:14.984547 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc8m7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46de52ee-226c-4713-bd7b-961f8f43f635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pj2t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc8m7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:14Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.003490 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:15Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.023024 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:15Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.038977 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:15Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.049585 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.049642 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.049655 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.049678 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.049693 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:15Z","lastTransitionTime":"2025-10-04T03:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.081407 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53d71d0eb0f06d88c47adfbd6bcb96a5c47a273919690b3cae9be51fff8b5a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f556ea9ccf10b49a8b1097267b402b532f8df5b1f47d7bcd0a45eb56eeaaedfc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:11:08Z\\\",\\\"message\\\":\\\"8.183889 6051 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:11:08.184000 6051 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:11:08.184057 6051 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:11:08.184239 6051 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:11:08.184529 6051 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:11:08.184927 6051 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:11:08.185386 6051 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 03:11:08.185496 6051 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:11:08.185515 6051 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 03:11:08.185549 6051 factory.go:656] Stopping watch factory\\\\nI1004 03:11:08.185554 6051 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 03:11:08.185557 6051 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:11:08.185600 6051 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d71d0eb0f06d88c47adfbd6bcb96a5c47a273919690b3cae9be51fff8b5a67\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:11:09Z\\\",\\\"message\\\":\\\"nd\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/console\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console/console_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/console\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.194\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1004 03:11:09.936938 6184 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Inte\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qrtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:15Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.106947 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://261e8ace3d214aec7b737ea538ac0996cc11a5dae65eacc33dd3a3ea65ed3558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:15Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.133493 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd1d6092975dd4703ff4930e2eafaf464f221510fddeda0f977a73c548c305eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:15Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.152323 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.152395 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.152406 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.152421 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.152433 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:15Z","lastTransitionTime":"2025-10-04T03:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.153110 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cwm6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a5aee9-b917-473a-9196-5cde6553c72a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6cefd4be7726c02ec9a7b2db28ce61c5dd11b708a131e05e96595b51719209b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrdmr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cwm6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:15Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.167084 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cd46x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0302950c-deb3-420a-b798-c99459e62b3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fls9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fls9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:12Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cd46x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:15Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.185785 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:15Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.199593 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11b178c3523269e8b4c09a1ad6f86eaee1d8f19f5056840dfa24466248889a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:15Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.257128 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.257196 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.257206 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.257226 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.257238 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:15Z","lastTransitionTime":"2025-10-04T03:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.361091 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.361180 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.361203 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.361237 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.361256 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:15Z","lastTransitionTime":"2025-10-04T03:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.464229 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.464322 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.464336 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.464360 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.464377 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:15Z","lastTransitionTime":"2025-10-04T03:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.567409 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.567476 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.567486 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.567505 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.567520 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:15Z","lastTransitionTime":"2025-10-04T03:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.670660 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.670724 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.670735 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.670771 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.670786 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:15Z","lastTransitionTime":"2025-10-04T03:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.774190 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.774260 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.774297 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.774314 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.774329 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:15Z","lastTransitionTime":"2025-10-04T03:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.858077 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:11:15 crc kubenswrapper[4742]: E1004 03:11:15.858458 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.878949 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.879014 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.879031 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.879056 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.879073 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:15Z","lastTransitionTime":"2025-10-04T03:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.983115 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.983213 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.983233 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.983332 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:15 crc kubenswrapper[4742]: I1004 03:11:15.983361 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:15Z","lastTransitionTime":"2025-10-04T03:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.086725 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.086781 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.086790 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.086809 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.086821 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:16Z","lastTransitionTime":"2025-10-04T03:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.189818 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.189874 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.189886 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.189906 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.189920 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:16Z","lastTransitionTime":"2025-10-04T03:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.293078 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.293134 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.293144 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.293163 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.293174 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:16Z","lastTransitionTime":"2025-10-04T03:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.372968 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0302950c-deb3-420a-b798-c99459e62b3f-metrics-certs\") pod \"network-metrics-daemon-cd46x\" (UID: \"0302950c-deb3-420a-b798-c99459e62b3f\") " pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:11:16 crc kubenswrapper[4742]: E1004 03:11:16.373217 4742 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:11:16 crc kubenswrapper[4742]: E1004 03:11:16.373371 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0302950c-deb3-420a-b798-c99459e62b3f-metrics-certs podName:0302950c-deb3-420a-b798-c99459e62b3f nodeName:}" failed. No retries permitted until 2025-10-04 03:11:20.373342938 +0000 UTC m=+46.305680284 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0302950c-deb3-420a-b798-c99459e62b3f-metrics-certs") pod "network-metrics-daemon-cd46x" (UID: "0302950c-deb3-420a-b798-c99459e62b3f") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.396013 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.396065 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.396077 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.396099 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.396113 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:16Z","lastTransitionTime":"2025-10-04T03:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.499494 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.499576 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.499603 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.499634 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.499658 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:16Z","lastTransitionTime":"2025-10-04T03:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.603500 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.603567 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.603590 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.603624 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.603651 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:16Z","lastTransitionTime":"2025-10-04T03:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.707256 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.707325 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.707337 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.707357 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.707370 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:16Z","lastTransitionTime":"2025-10-04T03:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.810266 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.810416 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.810438 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.810466 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.810484 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:16Z","lastTransitionTime":"2025-10-04T03:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.857529 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.857559 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.857630 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:16 crc kubenswrapper[4742]: E1004 03:11:16.857779 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:11:16 crc kubenswrapper[4742]: E1004 03:11:16.857873 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:11:16 crc kubenswrapper[4742]: E1004 03:11:16.857946 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.912489 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.912548 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.912565 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.912588 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:16 crc kubenswrapper[4742]: I1004 03:11:16.912606 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:16Z","lastTransitionTime":"2025-10-04T03:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.016537 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.016623 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.016639 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.016660 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.016674 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:17Z","lastTransitionTime":"2025-10-04T03:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.119310 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.119375 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.119394 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.119420 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.119437 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:17Z","lastTransitionTime":"2025-10-04T03:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.222253 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.222314 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.222330 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.222357 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.222371 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:17Z","lastTransitionTime":"2025-10-04T03:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.325177 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.325226 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.325237 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.325262 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.325301 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:17Z","lastTransitionTime":"2025-10-04T03:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.427490 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.427530 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.427540 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.427556 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.427566 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:17Z","lastTransitionTime":"2025-10-04T03:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.530450 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.530552 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.530564 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.530580 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.530589 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:17Z","lastTransitionTime":"2025-10-04T03:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.633235 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.633298 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.633308 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.633325 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.633340 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:17Z","lastTransitionTime":"2025-10-04T03:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.736533 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.736594 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.736607 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.736630 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.736644 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:17Z","lastTransitionTime":"2025-10-04T03:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.839849 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.840051 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.840132 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.840172 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.840244 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:17Z","lastTransitionTime":"2025-10-04T03:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.858210 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:11:17 crc kubenswrapper[4742]: E1004 03:11:17.858431 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.944210 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.944313 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.944339 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.944379 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:17 crc kubenswrapper[4742]: I1004 03:11:17.944404 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:17Z","lastTransitionTime":"2025-10-04T03:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.047528 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.047602 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.047624 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.047655 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.047690 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:18Z","lastTransitionTime":"2025-10-04T03:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.150786 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.150835 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.150846 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.150863 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.150875 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:18Z","lastTransitionTime":"2025-10-04T03:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.255504 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.255611 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.255640 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.255711 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.255750 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:18Z","lastTransitionTime":"2025-10-04T03:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.359554 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.359627 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.359650 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.359681 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.359702 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:18Z","lastTransitionTime":"2025-10-04T03:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.463180 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.463217 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.463226 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.463241 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.463250 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:18Z","lastTransitionTime":"2025-10-04T03:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.565859 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.565928 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.565939 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.565955 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.565966 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:18Z","lastTransitionTime":"2025-10-04T03:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.669556 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.669653 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.669685 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.669717 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.669737 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:18Z","lastTransitionTime":"2025-10-04T03:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.772327 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.772381 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.772399 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.772420 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.772434 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:18Z","lastTransitionTime":"2025-10-04T03:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.857810 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:11:18 crc kubenswrapper[4742]: E1004 03:11:18.857995 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.858090 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.858175 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:18 crc kubenswrapper[4742]: E1004 03:11:18.858414 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:11:18 crc kubenswrapper[4742]: E1004 03:11:18.858553 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.874655 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.874704 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.874717 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.874738 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.874751 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:18Z","lastTransitionTime":"2025-10-04T03:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.977266 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.977376 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.977396 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.977422 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:18 crc kubenswrapper[4742]: I1004 03:11:18.977441 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:18Z","lastTransitionTime":"2025-10-04T03:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.080644 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.080700 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.080712 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.080730 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.080744 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:19Z","lastTransitionTime":"2025-10-04T03:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.183148 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.183205 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.183217 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.183239 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.183254 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:19Z","lastTransitionTime":"2025-10-04T03:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.285784 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.285818 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.285828 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.285844 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.285857 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:19Z","lastTransitionTime":"2025-10-04T03:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.389410 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.389449 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.389460 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.389477 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.389489 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:19Z","lastTransitionTime":"2025-10-04T03:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.497151 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.497254 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.497379 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.497415 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.497439 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:19Z","lastTransitionTime":"2025-10-04T03:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.600785 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.600823 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.600833 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.600851 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.600862 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:19Z","lastTransitionTime":"2025-10-04T03:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.702807 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.702887 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.702904 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.702930 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.702949 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:19Z","lastTransitionTime":"2025-10-04T03:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.806230 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.806307 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.806318 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.806337 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.806348 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:19Z","lastTransitionTime":"2025-10-04T03:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.858269 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:11:19 crc kubenswrapper[4742]: E1004 03:11:19.858559 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.909208 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.909243 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.909253 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.909271 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:19 crc kubenswrapper[4742]: I1004 03:11:19.909300 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:19Z","lastTransitionTime":"2025-10-04T03:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.012692 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.012737 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.012748 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.012765 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.012779 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:20Z","lastTransitionTime":"2025-10-04T03:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.115203 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.115272 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.115317 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.115343 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.115362 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:20Z","lastTransitionTime":"2025-10-04T03:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.218461 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.218505 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.218520 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.218539 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.218551 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:20Z","lastTransitionTime":"2025-10-04T03:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.321479 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.321543 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.321555 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.321580 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.321594 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:20Z","lastTransitionTime":"2025-10-04T03:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.418664 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0302950c-deb3-420a-b798-c99459e62b3f-metrics-certs\") pod \"network-metrics-daemon-cd46x\" (UID: \"0302950c-deb3-420a-b798-c99459e62b3f\") " pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:11:20 crc kubenswrapper[4742]: E1004 03:11:20.418954 4742 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:11:20 crc kubenswrapper[4742]: E1004 03:11:20.419087 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0302950c-deb3-420a-b798-c99459e62b3f-metrics-certs podName:0302950c-deb3-420a-b798-c99459e62b3f nodeName:}" failed. No retries permitted until 2025-10-04 03:11:28.419060908 +0000 UTC m=+54.351398084 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0302950c-deb3-420a-b798-c99459e62b3f-metrics-certs") pod "network-metrics-daemon-cd46x" (UID: "0302950c-deb3-420a-b798-c99459e62b3f") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.424117 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.424187 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.424206 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.424234 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.424253 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:20Z","lastTransitionTime":"2025-10-04T03:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.533327 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.533400 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.533421 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.533451 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.533474 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:20Z","lastTransitionTime":"2025-10-04T03:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.637182 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.637238 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.637251 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.637293 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.637309 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:20Z","lastTransitionTime":"2025-10-04T03:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.740804 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.740890 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.740915 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.740949 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.740976 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:20Z","lastTransitionTime":"2025-10-04T03:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.844853 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.844929 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.844953 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.844985 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.845025 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:20Z","lastTransitionTime":"2025-10-04T03:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.858239 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.858296 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.858652 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:11:20 crc kubenswrapper[4742]: E1004 03:11:20.858649 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:11:20 crc kubenswrapper[4742]: E1004 03:11:20.858776 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:11:20 crc kubenswrapper[4742]: E1004 03:11:20.858905 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.859862 4742 scope.go:117] "RemoveContainer" containerID="53d71d0eb0f06d88c47adfbd6bcb96a5c47a273919690b3cae9be51fff8b5a67" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.885933 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.900831 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.902910 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.902989 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.903009 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.903063 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.903082 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:20Z","lastTransitionTime":"2025-10-04T03:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:20 crc kubenswrapper[4742]: E1004 03:11:20.921951 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.926258 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.928343 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.928378 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.928392 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.928414 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.928450 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:20Z","lastTransitionTime":"2025-10-04T03:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.942193 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:20 crc kubenswrapper[4742]: E1004 03:11:20.943865 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.949974 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.950008 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.950018 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.950036 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.950049 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:20Z","lastTransitionTime":"2025-10-04T03:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.962858 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.962922 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc8m7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46de52ee-226c-4713-bd7b-961f8f43f635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pj2t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc8m7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:20 crc kubenswrapper[4742]: E1004 03:11:20.968546 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.974947 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.974984 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.974997 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.975016 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.975029 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:20Z","lastTransitionTime":"2025-10-04T03:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.981835 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2694d67-7fb5-446f-a590-6113e9131921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b29154af28bbb704eeb5fde7a5c44db624f1cb35d79942c62d5ca9a03e697b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1070bde4d35430f4c712d63ede34a428ec20a277ce0409dd3e69dead27c3f98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwzp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:20 crc kubenswrapper[4742]: E1004 03:11:20.988316 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.993007 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.993072 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.993090 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.993118 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.993135 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:20Z","lastTransitionTime":"2025-10-04T03:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:20 crc kubenswrapper[4742]: I1004 03:11:20.998812 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:21 crc kubenswrapper[4742]: E1004 03:11:21.009594 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:21 crc kubenswrapper[4742]: E1004 03:11:21.009811 4742 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.013102 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.013174 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.013211 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.013242 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.013262 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:21Z","lastTransitionTime":"2025-10-04T03:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.017424 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.034488 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.063846 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53d71d0eb0f06d88c47adfbd6bcb96a5c47a273919690b3cae9be51fff8b5a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d71d0eb0f06d88c47adfbd6bcb96a5c47a273919690b3cae9be51fff8b5a67\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:11:09Z\\\",\\\"message\\\":\\\"nd\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/console\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console/console_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/console\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.194\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1004 03:11:09.936938 6184 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Inte\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-9qrtl_openshift-ovn-kubernetes(0cfc6e64-ba08-4c44-a18e-3b93ab792a6a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qrtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.079389 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://261e8ace3d214aec7b737ea538ac0996cc11a5dae65eacc33dd3a3ea65ed3558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.099927 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd1d6092975dd4703ff4930e2eafaf464f221510fddeda0f977a73c548c305eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.113362 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cwm6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a5aee9-b917-473a-9196-5cde6553c72a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6cefd4be7726c02ec9a7b2db28ce61c5dd11b708a131e05e96595b51719209b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrdmr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cwm6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.117383 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.117415 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.117425 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.117516 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.117532 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:21Z","lastTransitionTime":"2025-10-04T03:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.125728 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cd46x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0302950c-deb3-420a-b798-c99459e62b3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fls9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fls9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:12Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cd46x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.143798 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.156558 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11b178c3523269e8b4c09a1ad6f86eaee1d8f19f5056840dfa24466248889a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.171475 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qrtl_0cfc6e64-ba08-4c44-a18e-3b93ab792a6a/ovnkube-controller/1.log" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.175186 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" event={"ID":"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a","Type":"ContainerStarted","Data":"12a2784adf6efd5697d0afb2d3a66ea54c2b234d802f0065d6839be9d91fa920"} Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.175823 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.195565 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.214002 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.220182 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.220268 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.220344 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.220690 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.220866 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:21Z","lastTransitionTime":"2025-10-04T03:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.230815 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.254761 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.267908 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc8m7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46de52ee-226c-4713-bd7b-961f8f43f635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pj2t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc8m7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.280179 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2694d67-7fb5-446f-a590-6113e9131921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b29154af28bbb704eeb5fde7a5c44db624f1cb35d79942c62d5ca9a03e697b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1070bde4d35430f4c712d63ede34a428ec20a277ce0409dd3e69dead27c3f98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwzp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.302193 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.317861 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.324145 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.324190 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.324203 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.324222 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.324247 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:21Z","lastTransitionTime":"2025-10-04T03:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.339328 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.362090 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12a2784adf6efd5697d0afb2d3a66ea54c2b234d802f0065d6839be9d91fa920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d71d0eb0f06d88c47adfbd6bcb96a5c47a273919690b3cae9be51fff8b5a67\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:11:09Z\\\",\\\"message\\\":\\\"nd\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/console\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console/console_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/console\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.194\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1004 03:11:09.936938 6184 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Inte\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qrtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.378951 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://261e8ace3d214aec7b737ea538ac0996cc11a5dae65eacc33dd3a3ea65ed3558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.399857 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd1d6092975dd4703ff4930e2eafaf464f221510fddeda0f977a73c548c305eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.419545 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cwm6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a5aee9-b917-473a-9196-5cde6553c72a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6cefd4be7726c02ec9a7b2db28ce61c5dd11b708a131e05e96595b51719209b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrdmr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cwm6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.427372 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.427417 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.427447 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.427468 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.427480 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:21Z","lastTransitionTime":"2025-10-04T03:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.434270 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cd46x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0302950c-deb3-420a-b798-c99459e62b3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fls9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fls9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:12Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cd46x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.449166 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.461576 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11b178c3523269e8b4c09a1ad6f86eaee1d8f19f5056840dfa24466248889a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.529655 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.529716 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.529729 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.529749 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.529771 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:21Z","lastTransitionTime":"2025-10-04T03:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.632742 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.633022 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.633150 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.633270 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.633366 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:21Z","lastTransitionTime":"2025-10-04T03:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.736239 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.736318 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.736329 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.736350 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.736362 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:21Z","lastTransitionTime":"2025-10-04T03:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.838627 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.838714 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.838724 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.838741 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.838751 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:21Z","lastTransitionTime":"2025-10-04T03:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.858301 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:11:21 crc kubenswrapper[4742]: E1004 03:11:21.858500 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.941769 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.941823 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.941834 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.941852 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:21 crc kubenswrapper[4742]: I1004 03:11:21.941864 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:21Z","lastTransitionTime":"2025-10-04T03:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.044718 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.044757 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.044766 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.044782 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.044793 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:22Z","lastTransitionTime":"2025-10-04T03:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.147575 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.147659 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.147677 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.147704 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.147723 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:22Z","lastTransitionTime":"2025-10-04T03:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.182086 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qrtl_0cfc6e64-ba08-4c44-a18e-3b93ab792a6a/ovnkube-controller/2.log" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.182998 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qrtl_0cfc6e64-ba08-4c44-a18e-3b93ab792a6a/ovnkube-controller/1.log" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.186348 4742 generic.go:334] "Generic (PLEG): container finished" podID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerID="12a2784adf6efd5697d0afb2d3a66ea54c2b234d802f0065d6839be9d91fa920" exitCode=1 Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.186423 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" event={"ID":"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a","Type":"ContainerDied","Data":"12a2784adf6efd5697d0afb2d3a66ea54c2b234d802f0065d6839be9d91fa920"} Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.186503 4742 scope.go:117] "RemoveContainer" containerID="53d71d0eb0f06d88c47adfbd6bcb96a5c47a273919690b3cae9be51fff8b5a67" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.187337 4742 scope.go:117] "RemoveContainer" containerID="12a2784adf6efd5697d0afb2d3a66ea54c2b234d802f0065d6839be9d91fa920" Oct 04 03:11:22 crc kubenswrapper[4742]: E1004 03:11:22.187586 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9qrtl_openshift-ovn-kubernetes(0cfc6e64-ba08-4c44-a18e-3b93ab792a6a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.205808 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://261e8ace3d214aec7b737ea538ac0996cc11a5dae65eacc33dd3a3ea65ed3558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:22Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.221217 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd1d6092975dd4703ff4930e2eafaf464f221510fddeda0f977a73c548c305eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:22Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.236446 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cwm6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a5aee9-b917-473a-9196-5cde6553c72a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6cefd4be7726c02ec9a7b2db28ce61c5dd11b708a131e05e96595b51719209b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrdmr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cwm6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:22Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.250662 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.250646 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cd46x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0302950c-deb3-420a-b798-c99459e62b3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fls9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fls9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:12Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cd46x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:22Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.250721 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.250738 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.250763 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.250781 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:22Z","lastTransitionTime":"2025-10-04T03:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.270908 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:22Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.283722 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11b178c3523269e8b4c09a1ad6f86eaee1d8f19f5056840dfa24466248889a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:22Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.299679 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:22Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.316126 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:22Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.332133 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:22Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.350219 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc8m7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46de52ee-226c-4713-bd7b-961f8f43f635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pj2t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc8m7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:22Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.353049 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.353113 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.353128 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.353150 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.353163 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:22Z","lastTransitionTime":"2025-10-04T03:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.367297 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2694d67-7fb5-446f-a590-6113e9131921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b29154af28bbb704eeb5fde7a5c44db624f1cb35d79942c62d5ca9a03e697b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1070bde4d35430f4c712d63ede34a428ec20a277ce0409dd3e69dead27c3f98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwzp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:22Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.381530 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:22Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.397188 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:22Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.411525 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:22Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.425078 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:22Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.444794 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12a2784adf6efd5697d0afb2d3a66ea54c2b234d802f0065d6839be9d91fa920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d71d0eb0f06d88c47adfbd6bcb96a5c47a273919690b3cae9be51fff8b5a67\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:11:09Z\\\",\\\"message\\\":\\\"nd\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/console\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console/console_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/console\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.194\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1004 03:11:09.936938 6184 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Inte\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12a2784adf6efd5697d0afb2d3a66ea54c2b234d802f0065d6839be9d91fa920\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:11:21Z\\\",\\\"message\\\":\\\"0-73dcf8201ba5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 03:11:21.773108 6407 services_controller.go:451] Built service openshift-kube-storage-version-migrator-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1004 03:11:21.773205 6407 services_controller.go:452] Built service openshift-kube-storage-version-migrator-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1004 03:11:21.773214 6407 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/mac\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qrtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:22Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.456040 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.456122 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.456136 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.456156 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.456169 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:22Z","lastTransitionTime":"2025-10-04T03:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.559771 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.559848 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.559868 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.559906 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.559927 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:22Z","lastTransitionTime":"2025-10-04T03:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.662213 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.662259 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.662307 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.662328 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.662336 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:22Z","lastTransitionTime":"2025-10-04T03:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.765472 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.765586 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.765624 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.765658 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.765683 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:22Z","lastTransitionTime":"2025-10-04T03:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.857937 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.858023 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.858224 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:11:22 crc kubenswrapper[4742]: E1004 03:11:22.858217 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:11:22 crc kubenswrapper[4742]: E1004 03:11:22.858483 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:11:22 crc kubenswrapper[4742]: E1004 03:11:22.858587 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.868649 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.868715 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.868738 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.868767 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.868787 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:22Z","lastTransitionTime":"2025-10-04T03:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.971621 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.971688 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.971702 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.971724 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:22 crc kubenswrapper[4742]: I1004 03:11:22.971739 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:22Z","lastTransitionTime":"2025-10-04T03:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.079707 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.079788 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.079800 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.079820 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.079831 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:23Z","lastTransitionTime":"2025-10-04T03:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.183791 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.183867 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.183892 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.183927 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.183952 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:23Z","lastTransitionTime":"2025-10-04T03:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.193676 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qrtl_0cfc6e64-ba08-4c44-a18e-3b93ab792a6a/ovnkube-controller/2.log" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.198582 4742 scope.go:117] "RemoveContainer" containerID="12a2784adf6efd5697d0afb2d3a66ea54c2b234d802f0065d6839be9d91fa920" Oct 04 03:11:23 crc kubenswrapper[4742]: E1004 03:11:23.198867 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9qrtl_openshift-ovn-kubernetes(0cfc6e64-ba08-4c44-a18e-3b93ab792a6a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.224442 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:23Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.242335 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11b178c3523269e8b4c09a1ad6f86eaee1d8f19f5056840dfa24466248889a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:23Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.262009 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:23Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.276443 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:23Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.287216 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.287347 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.287379 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.287414 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.287440 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:23Z","lastTransitionTime":"2025-10-04T03:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.298144 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:23Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.314707 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc8m7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46de52ee-226c-4713-bd7b-961f8f43f635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pj2t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc8m7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:23Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.332251 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2694d67-7fb5-446f-a590-6113e9131921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b29154af28bbb704eeb5fde7a5c44db624f1cb35d79942c62d5ca9a03e697b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1070bde4d35430f4c712d63ede34a428ec20a277ce0409dd3e69dead27c3f98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwzp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:23Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.351967 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:23Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.366556 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:23Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.382828 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:23Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.390409 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.390459 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.390474 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.390496 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.390510 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:23Z","lastTransitionTime":"2025-10-04T03:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.398170 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:23Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.417935 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12a2784adf6efd5697d0afb2d3a66ea54c2b234d802f0065d6839be9d91fa920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12a2784adf6efd5697d0afb2d3a66ea54c2b234d802f0065d6839be9d91fa920\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:11:21Z\\\",\\\"message\\\":\\\"0-73dcf8201ba5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 03:11:21.773108 6407 services_controller.go:451] Built service openshift-kube-storage-version-migrator-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1004 03:11:21.773205 6407 services_controller.go:452] Built service openshift-kube-storage-version-migrator-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1004 03:11:21.773214 6407 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/mac\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9qrtl_openshift-ovn-kubernetes(0cfc6e64-ba08-4c44-a18e-3b93ab792a6a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qrtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:23Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.430906 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://261e8ace3d214aec7b737ea538ac0996cc11a5dae65eacc33dd3a3ea65ed3558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:23Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.445030 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd1d6092975dd4703ff4930e2eafaf464f221510fddeda0f977a73c548c305eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:23Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.456174 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cwm6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a5aee9-b917-473a-9196-5cde6553c72a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6cefd4be7726c02ec9a7b2db28ce61c5dd11b708a131e05e96595b51719209b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrdmr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cwm6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:23Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.465203 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cd46x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0302950c-deb3-420a-b798-c99459e62b3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fls9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fls9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:12Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cd46x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:23Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.493103 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.493162 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.493173 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.493194 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.493207 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:23Z","lastTransitionTime":"2025-10-04T03:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.595370 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.595412 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.595423 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.595445 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.595456 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:23Z","lastTransitionTime":"2025-10-04T03:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.698604 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.698646 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.698657 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.698674 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.698686 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:23Z","lastTransitionTime":"2025-10-04T03:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.800995 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.801061 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.801074 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.801094 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.801106 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:23Z","lastTransitionTime":"2025-10-04T03:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.857601 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:11:23 crc kubenswrapper[4742]: E1004 03:11:23.857835 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.903776 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.903823 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.903833 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.903852 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:23 crc kubenswrapper[4742]: I1004 03:11:23.903861 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:23Z","lastTransitionTime":"2025-10-04T03:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.006888 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.006928 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.006937 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.006952 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.006962 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:24Z","lastTransitionTime":"2025-10-04T03:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.109585 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.109631 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.109657 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.109674 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.109685 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:24Z","lastTransitionTime":"2025-10-04T03:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.212130 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.212208 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.212222 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.212246 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.212260 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:24Z","lastTransitionTime":"2025-10-04T03:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.314414 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.314482 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.314496 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.314516 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.314528 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:24Z","lastTransitionTime":"2025-10-04T03:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.417132 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.417180 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.417191 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.417208 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.417218 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:24Z","lastTransitionTime":"2025-10-04T03:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.520496 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.520547 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.520560 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.520574 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.520584 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:24Z","lastTransitionTime":"2025-10-04T03:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.623374 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.623435 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.623446 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.623466 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.623478 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:24Z","lastTransitionTime":"2025-10-04T03:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.726039 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.726089 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.726098 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.726119 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.726129 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:24Z","lastTransitionTime":"2025-10-04T03:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.811234 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.824943 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.828347 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.828411 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.828424 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.828440 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.828452 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:24Z","lastTransitionTime":"2025-10-04T03:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.829893 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc8m7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46de52ee-226c-4713-bd7b-961f8f43f635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pj2t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc8m7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:24Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.843433 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2694d67-7fb5-446f-a590-6113e9131921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b29154af28bbb704eeb5fde7a5c44db624f1cb35d79942c62d5ca9a03e697b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1070bde4d35430f4c712d63ede34a428ec20a277ce0409dd3e69dead27c3f98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwzp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:24Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.857851 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:24Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.858017 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.858119 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:24 crc kubenswrapper[4742]: E1004 03:11:24.858211 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.858251 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:11:24 crc kubenswrapper[4742]: E1004 03:11:24.858381 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:11:24 crc kubenswrapper[4742]: E1004 03:11:24.858447 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.873490 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:24Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.887161 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:24Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.904509 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:24Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.927148 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12a2784adf6efd5697d0afb2d3a66ea54c2b234d802f0065d6839be9d91fa920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12a2784adf6efd5697d0afb2d3a66ea54c2b234d802f0065d6839be9d91fa920\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:11:21Z\\\",\\\"message\\\":\\\"0-73dcf8201ba5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 03:11:21.773108 6407 services_controller.go:451] Built service openshift-kube-storage-version-migrator-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1004 03:11:21.773205 6407 services_controller.go:452] Built service openshift-kube-storage-version-migrator-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1004 03:11:21.773214 6407 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/mac\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9qrtl_openshift-ovn-kubernetes(0cfc6e64-ba08-4c44-a18e-3b93ab792a6a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qrtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:24Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.931323 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.931398 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.931410 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.931434 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.931455 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:24Z","lastTransitionTime":"2025-10-04T03:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.941663 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:24Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.961746 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:24Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.977540 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:24Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:24 crc kubenswrapper[4742]: I1004 03:11:24.988268 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cd46x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0302950c-deb3-420a-b798-c99459e62b3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fls9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fls9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:12Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cd46x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:24Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:24.999972 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://261e8ace3d214aec7b737ea538ac0996cc11a5dae65eacc33dd3a3ea65ed3558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:24Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.024162 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd1d6092975dd4703ff4930e2eafaf464f221510fddeda0f977a73c548c305eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:25Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.034556 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.034611 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.034624 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.034648 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.034665 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:25Z","lastTransitionTime":"2025-10-04T03:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.042539 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cwm6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a5aee9-b917-473a-9196-5cde6553c72a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6cefd4be7726c02ec9a7b2db28ce61c5dd11b708a131e05e96595b51719209b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrdmr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cwm6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:25Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.059199 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:25Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.071241 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11b178c3523269e8b4c09a1ad6f86eaee1d8f19f5056840dfa24466248889a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:25Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.085402 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cd46x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0302950c-deb3-420a-b798-c99459e62b3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fls9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fls9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:12Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cd46x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:25Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.098403 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://261e8ace3d214aec7b737ea538ac0996cc11a5dae65eacc33dd3a3ea65ed3558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:25Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.123296 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd1d6092975dd4703ff4930e2eafaf464f221510fddeda0f977a73c548c305eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:25Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.135600 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cwm6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a5aee9-b917-473a-9196-5cde6553c72a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6cefd4be7726c02ec9a7b2db28ce61c5dd11b708a131e05e96595b51719209b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrdmr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cwm6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:25Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.136934 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.136990 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.137002 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.137020 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.137033 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:25Z","lastTransitionTime":"2025-10-04T03:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.155984 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ec6cadd-fdb4-43c5-a898-f1bde3e9898d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf178a15abcb94eda5033b74ef132d9e204b33f4112152edf172e61999fc2c9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac618416ef312d2c051d88c0348fea230d73d3c8d6f4b1830b4b584ceae645c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2ede8f31552e98a3087393a27a4bab2c56f116829fef768b9c1dda8d6630142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3ee868ef16effadb77812f1673237d65d10493bd29dfcef5c7fe53da640d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce3ee868ef16effadb77812f1673237d65d10493bd29dfcef5c7fe53da640d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:25Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.178586 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:25Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.190517 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11b178c3523269e8b4c09a1ad6f86eaee1d8f19f5056840dfa24466248889a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:25Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.206295 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc8m7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46de52ee-226c-4713-bd7b-961f8f43f635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pj2t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc8m7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:25Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.222240 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2694d67-7fb5-446f-a590-6113e9131921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b29154af28bbb704eeb5fde7a5c44db624f1cb35d79942c62d5ca9a03e697b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1070bde4d35430f4c712d63ede34a428ec20a277ce0409dd3e69dead27c3f98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwzp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:25Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.241213 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.241443 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.241570 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.241666 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.241750 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:25Z","lastTransitionTime":"2025-10-04T03:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.242253 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:25Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.258071 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:25Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.274999 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:25Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.287731 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:25Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.309808 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12a2784adf6efd5697d0afb2d3a66ea54c2b234d802f0065d6839be9d91fa920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12a2784adf6efd5697d0afb2d3a66ea54c2b234d802f0065d6839be9d91fa920\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:11:21Z\\\",\\\"message\\\":\\\"0-73dcf8201ba5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 03:11:21.773108 6407 services_controller.go:451] Built service openshift-kube-storage-version-migrator-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1004 03:11:21.773205 6407 services_controller.go:452] Built service openshift-kube-storage-version-migrator-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1004 03:11:21.773214 6407 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/mac\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9qrtl_openshift-ovn-kubernetes(0cfc6e64-ba08-4c44-a18e-3b93ab792a6a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qrtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:25Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.326262 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:25Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.343456 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:25Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.344121 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.344156 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.344171 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.344199 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.344216 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:25Z","lastTransitionTime":"2025-10-04T03:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.359836 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:25Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.447000 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.447050 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.447061 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.447078 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.447091 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:25Z","lastTransitionTime":"2025-10-04T03:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.549946 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.549994 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.550008 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.550026 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.550038 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:25Z","lastTransitionTime":"2025-10-04T03:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.653200 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.653255 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.653267 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.653301 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.653312 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:25Z","lastTransitionTime":"2025-10-04T03:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.755964 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.756036 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.756050 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.756078 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.756105 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:25Z","lastTransitionTime":"2025-10-04T03:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.857580 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:11:25 crc kubenswrapper[4742]: E1004 03:11:25.857844 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.859599 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.859693 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.859707 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.859728 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.859743 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:25Z","lastTransitionTime":"2025-10-04T03:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.963356 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.963413 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.963423 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.963437 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:25 crc kubenswrapper[4742]: I1004 03:11:25.963447 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:25Z","lastTransitionTime":"2025-10-04T03:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.065735 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.065796 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.065808 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.065824 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.065835 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:26Z","lastTransitionTime":"2025-10-04T03:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.169111 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.169187 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.169199 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.169220 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.169236 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:26Z","lastTransitionTime":"2025-10-04T03:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.271496 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.271549 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.271561 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.271579 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.271591 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:26Z","lastTransitionTime":"2025-10-04T03:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.373886 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.373931 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.373941 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.373957 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.373969 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:26Z","lastTransitionTime":"2025-10-04T03:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.476426 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.476500 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.476513 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.476534 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.476546 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:26Z","lastTransitionTime":"2025-10-04T03:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.579443 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.579503 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.579513 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.579531 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.579542 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:26Z","lastTransitionTime":"2025-10-04T03:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.592080 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.592308 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:26 crc kubenswrapper[4742]: E1004 03:11:26.592424 4742 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:11:26 crc kubenswrapper[4742]: E1004 03:11:26.592458 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:11:58.592407031 +0000 UTC m=+84.524744257 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:11:26 crc kubenswrapper[4742]: E1004 03:11:26.592528 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:11:58.592505004 +0000 UTC m=+84.524842470 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.682906 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.683028 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.683050 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.683099 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.683111 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:26Z","lastTransitionTime":"2025-10-04T03:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.693164 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.693252 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.693365 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:11:26 crc kubenswrapper[4742]: E1004 03:11:26.693415 4742 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:11:26 crc kubenswrapper[4742]: E1004 03:11:26.693445 4742 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:11:26 crc kubenswrapper[4742]: E1004 03:11:26.693481 4742 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:11:26 crc kubenswrapper[4742]: E1004 03:11:26.693494 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:11:58.693472996 +0000 UTC m=+84.625810182 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:11:26 crc kubenswrapper[4742]: E1004 03:11:26.693502 4742 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:11:26 crc kubenswrapper[4742]: E1004 03:11:26.693567 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 03:11:58.693548828 +0000 UTC m=+84.625886044 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:11:26 crc kubenswrapper[4742]: E1004 03:11:26.693565 4742 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:11:26 crc kubenswrapper[4742]: E1004 03:11:26.693609 4742 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:11:26 crc kubenswrapper[4742]: E1004 03:11:26.693627 4742 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:11:26 crc kubenswrapper[4742]: E1004 03:11:26.693697 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 03:11:58.693679172 +0000 UTC m=+84.626016388 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.787396 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.787450 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.787462 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.787481 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.787494 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:26Z","lastTransitionTime":"2025-10-04T03:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.857737 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.857775 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:26 crc kubenswrapper[4742]: E1004 03:11:26.857945 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.857991 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:11:26 crc kubenswrapper[4742]: E1004 03:11:26.858197 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:11:26 crc kubenswrapper[4742]: E1004 03:11:26.858381 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.890501 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.890582 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.890605 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.890651 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.890681 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:26Z","lastTransitionTime":"2025-10-04T03:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.994239 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.994316 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.994329 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.994349 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:26 crc kubenswrapper[4742]: I1004 03:11:26.994364 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:26Z","lastTransitionTime":"2025-10-04T03:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.097050 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.097125 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.097147 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.097179 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.097202 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:27Z","lastTransitionTime":"2025-10-04T03:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.200519 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.200586 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.200603 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.200626 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.200643 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:27Z","lastTransitionTime":"2025-10-04T03:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.303769 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.303804 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.303813 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.303827 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.303838 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:27Z","lastTransitionTime":"2025-10-04T03:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.406696 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.406794 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.406822 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.406857 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.406882 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:27Z","lastTransitionTime":"2025-10-04T03:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.510344 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.510432 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.510452 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.510474 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.510487 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:27Z","lastTransitionTime":"2025-10-04T03:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.614129 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.614206 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.614232 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.614264 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.614328 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:27Z","lastTransitionTime":"2025-10-04T03:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.717701 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.717768 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.717779 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.717798 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.717810 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:27Z","lastTransitionTime":"2025-10-04T03:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.821482 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.821543 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.821554 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.821574 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.821586 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:27Z","lastTransitionTime":"2025-10-04T03:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.857994 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:11:27 crc kubenswrapper[4742]: E1004 03:11:27.858210 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.925369 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.925646 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.925717 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.925751 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:27 crc kubenswrapper[4742]: I1004 03:11:27.925771 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:27Z","lastTransitionTime":"2025-10-04T03:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.028674 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.028717 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.028727 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.028744 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.028754 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:28Z","lastTransitionTime":"2025-10-04T03:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.132004 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.132065 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.132080 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.132101 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.132114 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:28Z","lastTransitionTime":"2025-10-04T03:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.236366 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.236460 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.236477 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.236511 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.236538 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:28Z","lastTransitionTime":"2025-10-04T03:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.339489 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.339551 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.339579 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.339602 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.339614 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:28Z","lastTransitionTime":"2025-10-04T03:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.442869 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.442920 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.442931 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.442951 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.442963 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:28Z","lastTransitionTime":"2025-10-04T03:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.515165 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0302950c-deb3-420a-b798-c99459e62b3f-metrics-certs\") pod \"network-metrics-daemon-cd46x\" (UID: \"0302950c-deb3-420a-b798-c99459e62b3f\") " pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:11:28 crc kubenswrapper[4742]: E1004 03:11:28.515504 4742 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:11:28 crc kubenswrapper[4742]: E1004 03:11:28.515600 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0302950c-deb3-420a-b798-c99459e62b3f-metrics-certs podName:0302950c-deb3-420a-b798-c99459e62b3f nodeName:}" failed. No retries permitted until 2025-10-04 03:11:44.515575727 +0000 UTC m=+70.447912943 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0302950c-deb3-420a-b798-c99459e62b3f-metrics-certs") pod "network-metrics-daemon-cd46x" (UID: "0302950c-deb3-420a-b798-c99459e62b3f") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.545189 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.545249 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.545263 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.545309 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.545324 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:28Z","lastTransitionTime":"2025-10-04T03:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.647988 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.648029 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.648038 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.648054 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.648066 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:28Z","lastTransitionTime":"2025-10-04T03:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.750919 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.750986 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.750997 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.751013 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.751024 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:28Z","lastTransitionTime":"2025-10-04T03:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.853612 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.853669 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.853681 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.853702 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.853716 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:28Z","lastTransitionTime":"2025-10-04T03:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.857896 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.857945 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.857969 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:11:28 crc kubenswrapper[4742]: E1004 03:11:28.858034 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:11:28 crc kubenswrapper[4742]: E1004 03:11:28.858120 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:11:28 crc kubenswrapper[4742]: E1004 03:11:28.858238 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.956458 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.956506 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.956519 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.956535 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:28 crc kubenswrapper[4742]: I1004 03:11:28.956544 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:28Z","lastTransitionTime":"2025-10-04T03:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.059479 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.059537 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.059549 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.059569 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.059580 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:29Z","lastTransitionTime":"2025-10-04T03:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.162077 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.162127 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.162140 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.162158 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.162171 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:29Z","lastTransitionTime":"2025-10-04T03:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.265956 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.266161 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.266227 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.266354 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.266449 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:29Z","lastTransitionTime":"2025-10-04T03:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.368968 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.369039 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.369050 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.369072 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.369088 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:29Z","lastTransitionTime":"2025-10-04T03:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.472064 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.472120 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.472130 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.472150 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.472164 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:29Z","lastTransitionTime":"2025-10-04T03:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.574683 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.574739 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.574753 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.574773 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.574789 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:29Z","lastTransitionTime":"2025-10-04T03:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.677533 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.677568 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.677578 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.677591 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.677601 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:29Z","lastTransitionTime":"2025-10-04T03:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.780196 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.780233 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.780245 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.780262 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.780293 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:29Z","lastTransitionTime":"2025-10-04T03:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.858440 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:11:29 crc kubenswrapper[4742]: E1004 03:11:29.858602 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.882629 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.882666 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.882676 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.882692 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.882702 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:29Z","lastTransitionTime":"2025-10-04T03:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.984734 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.984775 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.984787 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.984803 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:29 crc kubenswrapper[4742]: I1004 03:11:29.984814 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:29Z","lastTransitionTime":"2025-10-04T03:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.086551 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.086585 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.086606 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.086621 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.086630 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:30Z","lastTransitionTime":"2025-10-04T03:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.189653 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.189685 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.189693 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.189711 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.189723 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:30Z","lastTransitionTime":"2025-10-04T03:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.292061 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.292093 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.292103 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.292121 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.292132 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:30Z","lastTransitionTime":"2025-10-04T03:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.394836 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.394897 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.394910 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.394928 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.394940 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:30Z","lastTransitionTime":"2025-10-04T03:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.497363 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.497397 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.497407 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.497424 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.497434 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:30Z","lastTransitionTime":"2025-10-04T03:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.600340 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.600383 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.600397 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.600416 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.600426 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:30Z","lastTransitionTime":"2025-10-04T03:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.703314 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.703376 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.703389 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.703408 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.703421 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:30Z","lastTransitionTime":"2025-10-04T03:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.806157 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.806199 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.806209 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.806224 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.806249 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:30Z","lastTransitionTime":"2025-10-04T03:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.858339 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.858942 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:11:30 crc kubenswrapper[4742]: E1004 03:11:30.859044 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:11:30 crc kubenswrapper[4742]: E1004 03:11:30.859083 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.859153 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:11:30 crc kubenswrapper[4742]: E1004 03:11:30.859208 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.908618 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.908685 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.908705 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.908733 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:30 crc kubenswrapper[4742]: I1004 03:11:30.908751 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:30Z","lastTransitionTime":"2025-10-04T03:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.011814 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.011896 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.011919 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.011948 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.011965 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:31Z","lastTransitionTime":"2025-10-04T03:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.114922 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.114976 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.114990 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.115011 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.115027 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:31Z","lastTransitionTime":"2025-10-04T03:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.217777 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.217873 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.217887 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.217910 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.217925 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:31Z","lastTransitionTime":"2025-10-04T03:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.321030 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.321088 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.321102 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.321140 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.321152 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:31Z","lastTransitionTime":"2025-10-04T03:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.347130 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.347178 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.347198 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.347221 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.347238 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:31Z","lastTransitionTime":"2025-10-04T03:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:31 crc kubenswrapper[4742]: E1004 03:11:31.363329 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:31Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.367700 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.367731 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.367744 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.367762 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.367775 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:31Z","lastTransitionTime":"2025-10-04T03:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:31 crc kubenswrapper[4742]: E1004 03:11:31.385363 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:31Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.390093 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.390176 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.390200 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.390236 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.390261 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:31Z","lastTransitionTime":"2025-10-04T03:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:31 crc kubenswrapper[4742]: E1004 03:11:31.406569 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:31Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.411104 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.411151 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.411172 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.411201 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.411221 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:31Z","lastTransitionTime":"2025-10-04T03:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:31 crc kubenswrapper[4742]: E1004 03:11:31.431575 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:31Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.436394 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.436438 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.436450 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.436468 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.436481 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:31Z","lastTransitionTime":"2025-10-04T03:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:31 crc kubenswrapper[4742]: E1004 03:11:31.453880 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:31Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:31 crc kubenswrapper[4742]: E1004 03:11:31.453995 4742 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.455787 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.455815 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.455826 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.455842 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.455854 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:31Z","lastTransitionTime":"2025-10-04T03:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.558318 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.558371 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.558384 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.558401 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.558413 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:31Z","lastTransitionTime":"2025-10-04T03:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.661298 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.661344 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.661353 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.661369 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.661379 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:31Z","lastTransitionTime":"2025-10-04T03:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.764219 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.764286 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.764298 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.764317 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.764328 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:31Z","lastTransitionTime":"2025-10-04T03:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.857664 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:11:31 crc kubenswrapper[4742]: E1004 03:11:31.857844 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.866568 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.866616 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.866631 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.866652 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.866666 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:31Z","lastTransitionTime":"2025-10-04T03:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.969665 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.969709 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.969719 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.969739 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:31 crc kubenswrapper[4742]: I1004 03:11:31.969751 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:31Z","lastTransitionTime":"2025-10-04T03:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.072172 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.072229 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.072247 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.072306 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.072325 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:32Z","lastTransitionTime":"2025-10-04T03:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.174976 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.175045 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.175060 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.175080 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.175092 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:32Z","lastTransitionTime":"2025-10-04T03:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.277778 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.277826 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.277838 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.277856 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.277865 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:32Z","lastTransitionTime":"2025-10-04T03:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.380357 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.380522 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.380578 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.380636 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.380648 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:32Z","lastTransitionTime":"2025-10-04T03:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.483080 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.483139 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.483149 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.483168 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.483179 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:32Z","lastTransitionTime":"2025-10-04T03:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.585797 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.585854 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.585875 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.585896 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.585909 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:32Z","lastTransitionTime":"2025-10-04T03:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.688216 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.688290 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.688304 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.688323 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.688334 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:32Z","lastTransitionTime":"2025-10-04T03:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.791124 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.791170 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.791182 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.791200 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.791213 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:32Z","lastTransitionTime":"2025-10-04T03:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.858466 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.858556 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.858666 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:32 crc kubenswrapper[4742]: E1004 03:11:32.858831 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:11:32 crc kubenswrapper[4742]: E1004 03:11:32.858927 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:11:32 crc kubenswrapper[4742]: E1004 03:11:32.859110 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.894330 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.894383 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.894396 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.894414 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.894427 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:32Z","lastTransitionTime":"2025-10-04T03:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.996881 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.996923 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.996932 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.996946 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:32 crc kubenswrapper[4742]: I1004 03:11:32.996957 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:32Z","lastTransitionTime":"2025-10-04T03:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.101021 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.101091 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.101104 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.101125 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.101148 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:33Z","lastTransitionTime":"2025-10-04T03:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.204120 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.204182 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.204196 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.204217 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.204232 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:33Z","lastTransitionTime":"2025-10-04T03:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.306479 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.306832 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.306842 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.306858 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.306868 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:33Z","lastTransitionTime":"2025-10-04T03:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.409032 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.409077 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.409086 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.409102 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.409113 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:33Z","lastTransitionTime":"2025-10-04T03:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.512343 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.512418 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.512430 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.512447 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.512460 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:33Z","lastTransitionTime":"2025-10-04T03:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.614675 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.614736 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.614754 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.614777 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.614794 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:33Z","lastTransitionTime":"2025-10-04T03:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.717737 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.717790 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.717814 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.717832 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.717844 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:33Z","lastTransitionTime":"2025-10-04T03:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.819674 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.819723 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.819737 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.819756 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.819777 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:33Z","lastTransitionTime":"2025-10-04T03:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.857469 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:11:33 crc kubenswrapper[4742]: E1004 03:11:33.857668 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.922884 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.922949 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.922966 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.922988 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:33 crc kubenswrapper[4742]: I1004 03:11:33.923002 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:33Z","lastTransitionTime":"2025-10-04T03:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.025698 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.025780 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.025803 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.025829 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.025847 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:34Z","lastTransitionTime":"2025-10-04T03:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.128133 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.128200 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.128212 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.128231 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.128242 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:34Z","lastTransitionTime":"2025-10-04T03:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.231800 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.232063 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.232084 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.232115 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.232145 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:34Z","lastTransitionTime":"2025-10-04T03:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.335365 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.335417 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.335432 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.335451 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.335465 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:34Z","lastTransitionTime":"2025-10-04T03:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.437872 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.437921 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.437934 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.437951 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.437962 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:34Z","lastTransitionTime":"2025-10-04T03:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.540708 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.540753 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.540764 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.540782 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.540792 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:34Z","lastTransitionTime":"2025-10-04T03:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.643060 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.643108 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.643117 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.643133 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.643143 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:34Z","lastTransitionTime":"2025-10-04T03:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.745510 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.745559 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.745572 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.745589 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.745603 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:34Z","lastTransitionTime":"2025-10-04T03:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.849077 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.849169 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.849198 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.849339 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.849383 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:34Z","lastTransitionTime":"2025-10-04T03:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.857413 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.857488 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.857513 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:11:34 crc kubenswrapper[4742]: E1004 03:11:34.857686 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:11:34 crc kubenswrapper[4742]: E1004 03:11:34.857907 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:11:34 crc kubenswrapper[4742]: E1004 03:11:34.858050 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.873890 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ec6cadd-fdb4-43c5-a898-f1bde3e9898d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf178a15abcb94eda5033b74ef132d9e204b33f4112152edf172e61999fc2c9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac618416ef312d2c051d88c0348fea230d73d3c8d6f4b1830b4b584ceae645c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2ede8f31552e98a3087393a27a4bab2c56f116829fef768b9c1dda8d6630142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3ee868ef16effadb77812f1673237d65d10493bd29dfcef5c7fe53da640d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce3ee868ef16effadb77812f1673237d65d10493bd29dfcef5c7fe53da640d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.887388 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.898640 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11b178c3523269e8b4c09a1ad6f86eaee1d8f19f5056840dfa24466248889a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.911134 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.923077 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.935577 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.947005 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.951728 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.951752 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.951760 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.951775 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.951784 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:34Z","lastTransitionTime":"2025-10-04T03:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.961825 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc8m7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46de52ee-226c-4713-bd7b-961f8f43f635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pj2t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc8m7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.974492 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2694d67-7fb5-446f-a590-6113e9131921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b29154af28bbb704eeb5fde7a5c44db624f1cb35d79942c62d5ca9a03e697b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1070bde4d35430f4c712d63ede34a428ec20a277ce0409dd3e69dead27c3f98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwzp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.986905 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:34 crc kubenswrapper[4742]: I1004 03:11:34.999068 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.010045 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:35Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.027237 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12a2784adf6efd5697d0afb2d3a66ea54c2b234d802f0065d6839be9d91fa920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12a2784adf6efd5697d0afb2d3a66ea54c2b234d802f0065d6839be9d91fa920\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:11:21Z\\\",\\\"message\\\":\\\"0-73dcf8201ba5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 03:11:21.773108 6407 services_controller.go:451] Built service openshift-kube-storage-version-migrator-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1004 03:11:21.773205 6407 services_controller.go:452] Built service openshift-kube-storage-version-migrator-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1004 03:11:21.773214 6407 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/mac\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9qrtl_openshift-ovn-kubernetes(0cfc6e64-ba08-4c44-a18e-3b93ab792a6a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qrtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:35Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.040180 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://261e8ace3d214aec7b737ea538ac0996cc11a5dae65eacc33dd3a3ea65ed3558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:35Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.054438 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.054597 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.054713 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.054828 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.054940 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:35Z","lastTransitionTime":"2025-10-04T03:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.055060 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd1d6092975dd4703ff4930e2eafaf464f221510fddeda0f977a73c548c305eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:35Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.067720 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cwm6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a5aee9-b917-473a-9196-5cde6553c72a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6cefd4be7726c02ec9a7b2db28ce61c5dd11b708a131e05e96595b51719209b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrdmr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cwm6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:35Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.077647 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cd46x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0302950c-deb3-420a-b798-c99459e62b3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fls9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fls9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:12Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cd46x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:35Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.158406 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.158457 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.158470 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.158487 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.158498 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:35Z","lastTransitionTime":"2025-10-04T03:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.261149 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.261225 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.261243 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.261294 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.261312 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:35Z","lastTransitionTime":"2025-10-04T03:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.363578 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.363623 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.363632 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.363651 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.363664 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:35Z","lastTransitionTime":"2025-10-04T03:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.466671 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.466725 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.466742 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.466765 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.466780 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:35Z","lastTransitionTime":"2025-10-04T03:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.569163 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.569221 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.569238 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.569259 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.569316 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:35Z","lastTransitionTime":"2025-10-04T03:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.672365 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.672517 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.672542 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.672615 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.672640 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:35Z","lastTransitionTime":"2025-10-04T03:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.776211 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.776295 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.776311 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.776334 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.776346 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:35Z","lastTransitionTime":"2025-10-04T03:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.857650 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:11:35 crc kubenswrapper[4742]: E1004 03:11:35.858089 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.879252 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.879341 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.879356 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.879380 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.879396 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:35Z","lastTransitionTime":"2025-10-04T03:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.982522 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.982575 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.982591 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.982612 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:35 crc kubenswrapper[4742]: I1004 03:11:35.982630 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:35Z","lastTransitionTime":"2025-10-04T03:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.085210 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.085255 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.085266 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.085301 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.085316 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:36Z","lastTransitionTime":"2025-10-04T03:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.188746 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.188788 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.188799 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.188815 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.188825 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:36Z","lastTransitionTime":"2025-10-04T03:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.291326 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.291374 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.291393 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.291413 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.291429 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:36Z","lastTransitionTime":"2025-10-04T03:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.393907 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.393948 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.393957 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.393970 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.393980 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:36Z","lastTransitionTime":"2025-10-04T03:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.496803 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.496844 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.496864 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.496878 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.496888 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:36Z","lastTransitionTime":"2025-10-04T03:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.599541 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.599582 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.599591 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.599607 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.599619 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:36Z","lastTransitionTime":"2025-10-04T03:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.702703 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.702742 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.702750 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.702765 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.702777 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:36Z","lastTransitionTime":"2025-10-04T03:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.805007 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.805052 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.805062 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.805080 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.805090 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:36Z","lastTransitionTime":"2025-10-04T03:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.858137 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.858251 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:36 crc kubenswrapper[4742]: E1004 03:11:36.858364 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.858416 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:11:36 crc kubenswrapper[4742]: E1004 03:11:36.858532 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:11:36 crc kubenswrapper[4742]: E1004 03:11:36.858634 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.907148 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.907203 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.907215 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.907236 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:36 crc kubenswrapper[4742]: I1004 03:11:36.907251 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:36Z","lastTransitionTime":"2025-10-04T03:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.009873 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.009923 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.009934 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.009951 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.009962 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:37Z","lastTransitionTime":"2025-10-04T03:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.112237 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.112291 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.112301 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.112316 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.112328 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:37Z","lastTransitionTime":"2025-10-04T03:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.214667 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.214707 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.214716 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.214730 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.214739 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:37Z","lastTransitionTime":"2025-10-04T03:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.316777 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.316813 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.316822 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.316835 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.316846 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:37Z","lastTransitionTime":"2025-10-04T03:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.418751 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.418785 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.418795 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.418808 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.418817 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:37Z","lastTransitionTime":"2025-10-04T03:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.521179 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.521221 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.521231 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.521245 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.521254 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:37Z","lastTransitionTime":"2025-10-04T03:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.624521 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.624586 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.624599 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.624624 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.624635 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:37Z","lastTransitionTime":"2025-10-04T03:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.726785 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.726817 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.726825 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.726839 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.726848 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:37Z","lastTransitionTime":"2025-10-04T03:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.829208 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.829252 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.829264 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.829302 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.829374 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:37Z","lastTransitionTime":"2025-10-04T03:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.857618 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:11:37 crc kubenswrapper[4742]: E1004 03:11:37.858013 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.858215 4742 scope.go:117] "RemoveContainer" containerID="12a2784adf6efd5697d0afb2d3a66ea54c2b234d802f0065d6839be9d91fa920" Oct 04 03:11:37 crc kubenswrapper[4742]: E1004 03:11:37.858461 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9qrtl_openshift-ovn-kubernetes(0cfc6e64-ba08-4c44-a18e-3b93ab792a6a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.932208 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.932252 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.932287 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.932309 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:37 crc kubenswrapper[4742]: I1004 03:11:37.932322 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:37Z","lastTransitionTime":"2025-10-04T03:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.034421 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.034489 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.034502 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.034520 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.034531 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:38Z","lastTransitionTime":"2025-10-04T03:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.136800 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.136840 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.136851 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.136872 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.136884 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:38Z","lastTransitionTime":"2025-10-04T03:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.239526 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.239578 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.239590 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.239608 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.239621 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:38Z","lastTransitionTime":"2025-10-04T03:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.341545 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.341582 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.341597 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.341614 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.341623 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:38Z","lastTransitionTime":"2025-10-04T03:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.444415 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.444461 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.444471 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.444490 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.444501 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:38Z","lastTransitionTime":"2025-10-04T03:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.547163 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.547253 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.547266 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.547316 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.547330 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:38Z","lastTransitionTime":"2025-10-04T03:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.649995 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.650044 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.650055 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.650072 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.650081 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:38Z","lastTransitionTime":"2025-10-04T03:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.751863 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.751922 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.751932 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.751948 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.751958 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:38Z","lastTransitionTime":"2025-10-04T03:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.855118 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.855172 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.855186 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.855205 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.855221 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:38Z","lastTransitionTime":"2025-10-04T03:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.857459 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.857514 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.857528 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:11:38 crc kubenswrapper[4742]: E1004 03:11:38.857576 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:11:38 crc kubenswrapper[4742]: E1004 03:11:38.857637 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:11:38 crc kubenswrapper[4742]: E1004 03:11:38.857738 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.957808 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.957853 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.957865 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.957883 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:38 crc kubenswrapper[4742]: I1004 03:11:38.957895 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:38Z","lastTransitionTime":"2025-10-04T03:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.060485 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.060547 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.060557 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.060573 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.060582 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:39Z","lastTransitionTime":"2025-10-04T03:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.163466 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.163526 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.163536 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.163556 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.163570 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:39Z","lastTransitionTime":"2025-10-04T03:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.266589 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.266632 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.266641 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.266660 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.266672 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:39Z","lastTransitionTime":"2025-10-04T03:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.369591 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.369635 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.369648 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.369669 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.369683 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:39Z","lastTransitionTime":"2025-10-04T03:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.472397 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.472428 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.472438 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.472454 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.472464 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:39Z","lastTransitionTime":"2025-10-04T03:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.574662 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.575078 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.575236 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.575439 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.575583 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:39Z","lastTransitionTime":"2025-10-04T03:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.679429 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.679471 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.679479 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.679497 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.679509 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:39Z","lastTransitionTime":"2025-10-04T03:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.782142 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.782459 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.782527 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.782606 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.782672 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:39Z","lastTransitionTime":"2025-10-04T03:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.858022 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:11:39 crc kubenswrapper[4742]: E1004 03:11:39.858187 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.885687 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.885735 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.885749 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.885809 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.885824 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:39Z","lastTransitionTime":"2025-10-04T03:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.988981 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.989057 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.989069 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.989092 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:39 crc kubenswrapper[4742]: I1004 03:11:39.989105 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:39Z","lastTransitionTime":"2025-10-04T03:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.091812 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.091859 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.091869 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.091886 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.091898 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:40Z","lastTransitionTime":"2025-10-04T03:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.195032 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.195083 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.195094 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.195114 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.195156 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:40Z","lastTransitionTime":"2025-10-04T03:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.297438 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.297496 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.297508 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.297527 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.297539 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:40Z","lastTransitionTime":"2025-10-04T03:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.400065 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.400111 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.400119 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.400136 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.400145 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:40Z","lastTransitionTime":"2025-10-04T03:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.503022 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.503104 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.503129 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.503195 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.503217 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:40Z","lastTransitionTime":"2025-10-04T03:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.605432 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.605494 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.605506 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.605523 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.605534 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:40Z","lastTransitionTime":"2025-10-04T03:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.707727 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.707765 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.707779 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.707794 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.707805 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:40Z","lastTransitionTime":"2025-10-04T03:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.812568 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.812612 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.812623 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.812642 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.812655 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:40Z","lastTransitionTime":"2025-10-04T03:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.857780 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.857814 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.857817 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:40 crc kubenswrapper[4742]: E1004 03:11:40.857968 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:11:40 crc kubenswrapper[4742]: E1004 03:11:40.858074 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:11:40 crc kubenswrapper[4742]: E1004 03:11:40.858241 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.915399 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.915443 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.915452 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.915468 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:40 crc kubenswrapper[4742]: I1004 03:11:40.915478 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:40Z","lastTransitionTime":"2025-10-04T03:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.018037 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.018085 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.018095 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.018110 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.018121 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:41Z","lastTransitionTime":"2025-10-04T03:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.120585 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.120635 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.120648 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.120666 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.120682 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:41Z","lastTransitionTime":"2025-10-04T03:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.223902 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.223944 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.223956 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.223970 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.223980 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:41Z","lastTransitionTime":"2025-10-04T03:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.326659 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.326708 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.326719 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.326738 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.326755 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:41Z","lastTransitionTime":"2025-10-04T03:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.429311 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.429361 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.429376 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.429397 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.429413 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:41Z","lastTransitionTime":"2025-10-04T03:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.531566 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.531609 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.531621 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.531636 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.531646 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:41Z","lastTransitionTime":"2025-10-04T03:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.599063 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.599348 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.599448 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.599554 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.599636 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:41Z","lastTransitionTime":"2025-10-04T03:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:41 crc kubenswrapper[4742]: E1004 03:11:41.612722 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:41Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.616876 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.616935 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.616946 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.616968 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.616979 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:41Z","lastTransitionTime":"2025-10-04T03:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:41 crc kubenswrapper[4742]: E1004 03:11:41.632030 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:41Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.636363 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.636401 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.636414 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.636432 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.636444 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:41Z","lastTransitionTime":"2025-10-04T03:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:41 crc kubenswrapper[4742]: E1004 03:11:41.649937 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:41Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.657951 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.658727 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.658748 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.658772 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.658789 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:41Z","lastTransitionTime":"2025-10-04T03:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:41 crc kubenswrapper[4742]: E1004 03:11:41.673058 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:41Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.677565 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.677610 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.677620 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.677640 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.677652 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:41Z","lastTransitionTime":"2025-10-04T03:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:41 crc kubenswrapper[4742]: E1004 03:11:41.692139 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:41Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:41 crc kubenswrapper[4742]: E1004 03:11:41.692264 4742 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.694335 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.694415 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.694426 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.694446 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.694462 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:41Z","lastTransitionTime":"2025-10-04T03:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.797297 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.797345 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.797356 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.797373 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.797384 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:41Z","lastTransitionTime":"2025-10-04T03:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.857980 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:11:41 crc kubenswrapper[4742]: E1004 03:11:41.858351 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.899837 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.899909 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.899939 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.899963 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:41 crc kubenswrapper[4742]: I1004 03:11:41.899977 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:41Z","lastTransitionTime":"2025-10-04T03:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.002908 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.002969 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.002980 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.002998 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.003015 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:42Z","lastTransitionTime":"2025-10-04T03:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.105720 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.105782 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.105793 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.105811 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.105824 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:42Z","lastTransitionTime":"2025-10-04T03:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.207919 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.207959 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.207970 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.207988 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.208001 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:42Z","lastTransitionTime":"2025-10-04T03:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.310106 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.310176 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.310186 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.310203 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.310216 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:42Z","lastTransitionTime":"2025-10-04T03:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.413044 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.413137 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.413148 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.413169 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.413182 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:42Z","lastTransitionTime":"2025-10-04T03:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.515720 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.515777 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.515789 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.515807 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.515820 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:42Z","lastTransitionTime":"2025-10-04T03:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.618138 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.618236 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.618255 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.618338 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.618425 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:42Z","lastTransitionTime":"2025-10-04T03:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.721682 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.721744 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.721758 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.721781 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.721802 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:42Z","lastTransitionTime":"2025-10-04T03:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.824252 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.824301 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.824311 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.824325 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.824335 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:42Z","lastTransitionTime":"2025-10-04T03:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.857886 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.857970 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:42 crc kubenswrapper[4742]: E1004 03:11:42.858095 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:11:42 crc kubenswrapper[4742]: E1004 03:11:42.858264 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.858443 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:11:42 crc kubenswrapper[4742]: E1004 03:11:42.858617 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.927450 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.927500 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.927509 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.927526 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:42 crc kubenswrapper[4742]: I1004 03:11:42.927535 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:42Z","lastTransitionTime":"2025-10-04T03:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.030574 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.030629 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.030639 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.030658 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.030670 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:43Z","lastTransitionTime":"2025-10-04T03:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.133459 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.133514 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.133525 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.133545 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.133558 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:43Z","lastTransitionTime":"2025-10-04T03:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.236023 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.236084 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.236096 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.236124 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.236142 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:43Z","lastTransitionTime":"2025-10-04T03:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.338354 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.338402 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.338412 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.338429 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.338440 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:43Z","lastTransitionTime":"2025-10-04T03:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.440851 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.440909 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.440920 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.440940 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.440951 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:43Z","lastTransitionTime":"2025-10-04T03:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.543602 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.543660 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.543670 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.543687 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.543698 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:43Z","lastTransitionTime":"2025-10-04T03:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.646906 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.646953 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.646963 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.646980 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.646991 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:43Z","lastTransitionTime":"2025-10-04T03:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.749944 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.750000 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.750014 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.750039 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.750055 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:43Z","lastTransitionTime":"2025-10-04T03:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.852990 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.853068 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.853090 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.853118 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.853136 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:43Z","lastTransitionTime":"2025-10-04T03:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.858188 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:11:43 crc kubenswrapper[4742]: E1004 03:11:43.858382 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.955979 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.956097 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.956112 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.956133 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:43 crc kubenswrapper[4742]: I1004 03:11:43.956162 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:43Z","lastTransitionTime":"2025-10-04T03:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.059216 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.059263 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.059325 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.059344 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.059357 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:44Z","lastTransitionTime":"2025-10-04T03:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.161891 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.161930 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.161948 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.161967 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.161979 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:44Z","lastTransitionTime":"2025-10-04T03:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.263815 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.263859 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.263868 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.263883 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.263896 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:44Z","lastTransitionTime":"2025-10-04T03:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.366323 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.366370 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.366382 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.366405 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.366420 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:44Z","lastTransitionTime":"2025-10-04T03:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.469529 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.469579 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.469590 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.469610 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.469620 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:44Z","lastTransitionTime":"2025-10-04T03:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.571797 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.571834 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.571845 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.571865 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.571878 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:44Z","lastTransitionTime":"2025-10-04T03:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.590722 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0302950c-deb3-420a-b798-c99459e62b3f-metrics-certs\") pod \"network-metrics-daemon-cd46x\" (UID: \"0302950c-deb3-420a-b798-c99459e62b3f\") " pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:11:44 crc kubenswrapper[4742]: E1004 03:11:44.590988 4742 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:11:44 crc kubenswrapper[4742]: E1004 03:11:44.591152 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0302950c-deb3-420a-b798-c99459e62b3f-metrics-certs podName:0302950c-deb3-420a-b798-c99459e62b3f nodeName:}" failed. No retries permitted until 2025-10-04 03:12:16.591093111 +0000 UTC m=+102.523430287 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0302950c-deb3-420a-b798-c99459e62b3f-metrics-certs") pod "network-metrics-daemon-cd46x" (UID: "0302950c-deb3-420a-b798-c99459e62b3f") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.674582 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.674641 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.674650 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.674667 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.674680 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:44Z","lastTransitionTime":"2025-10-04T03:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.777464 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.777556 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.777574 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.777592 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.777604 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:44Z","lastTransitionTime":"2025-10-04T03:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.857681 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.857724 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:11:44 crc kubenswrapper[4742]: E1004 03:11:44.857868 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:11:44 crc kubenswrapper[4742]: E1004 03:11:44.857995 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.858109 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:11:44 crc kubenswrapper[4742]: E1004 03:11:44.858203 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.871112 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://261e8ace3d214aec7b737ea538ac0996cc11a5dae65eacc33dd3a3ea65ed3558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:44Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.879543 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.879576 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.879585 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.879598 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.879609 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:44Z","lastTransitionTime":"2025-10-04T03:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.886663 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd1d6092975dd4703ff4930e2eafaf464f221510fddeda0f977a73c548c305eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:44Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.898818 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cwm6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a5aee9-b917-473a-9196-5cde6553c72a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6cefd4be7726c02ec9a7b2db28ce61c5dd11b708a131e05e96595b51719209b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrdmr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cwm6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:44Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.912257 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cd46x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0302950c-deb3-420a-b798-c99459e62b3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fls9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fls9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:12Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cd46x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:44Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.925092 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ec6cadd-fdb4-43c5-a898-f1bde3e9898d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf178a15abcb94eda5033b74ef132d9e204b33f4112152edf172e61999fc2c9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac618416ef312d2c051d88c0348fea230d73d3c8d6f4b1830b4b584ceae645c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2ede8f31552e98a3087393a27a4bab2c56f116829fef768b9c1dda8d6630142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3ee868ef16effadb77812f1673237d65d10493bd29dfcef5c7fe53da640d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce3ee868ef16effadb77812f1673237d65d10493bd29dfcef5c7fe53da640d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:44Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.941898 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:44Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.956247 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11b178c3523269e8b4c09a1ad6f86eaee1d8f19f5056840dfa24466248889a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:44Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.972237 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:44Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.982132 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.982194 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.982205 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.982222 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.982234 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:44Z","lastTransitionTime":"2025-10-04T03:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:44 crc kubenswrapper[4742]: I1004 03:11:44.988404 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:44Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.004699 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.021617 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.037294 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc8m7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46de52ee-226c-4713-bd7b-961f8f43f635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pj2t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc8m7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.052092 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2694d67-7fb5-446f-a590-6113e9131921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b29154af28bbb704eeb5fde7a5c44db624f1cb35d79942c62d5ca9a03e697b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1070bde4d35430f4c712d63ede34a428ec20a277ce0409dd3e69dead27c3f98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwzp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.066255 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.080244 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.084863 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.084922 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.084937 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.084962 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.084976 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:45Z","lastTransitionTime":"2025-10-04T03:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.092111 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.114302 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12a2784adf6efd5697d0afb2d3a66ea54c2b234d802f0065d6839be9d91fa920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12a2784adf6efd5697d0afb2d3a66ea54c2b234d802f0065d6839be9d91fa920\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:11:21Z\\\",\\\"message\\\":\\\"0-73dcf8201ba5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 03:11:21.773108 6407 services_controller.go:451] Built service openshift-kube-storage-version-migrator-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1004 03:11:21.773205 6407 services_controller.go:452] Built service openshift-kube-storage-version-migrator-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1004 03:11:21.773214 6407 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/mac\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9qrtl_openshift-ovn-kubernetes(0cfc6e64-ba08-4c44-a18e-3b93ab792a6a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qrtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.188081 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.188135 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.188148 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.188169 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.188185 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:45Z","lastTransitionTime":"2025-10-04T03:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.291813 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.291867 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.291878 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.291903 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.291919 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:45Z","lastTransitionTime":"2025-10-04T03:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.395167 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.395222 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.395235 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.395255 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.395289 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:45Z","lastTransitionTime":"2025-10-04T03:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.498775 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.498840 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.498862 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.498902 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.498925 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:45Z","lastTransitionTime":"2025-10-04T03:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.601756 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.601826 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.601840 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.601861 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.601876 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:45Z","lastTransitionTime":"2025-10-04T03:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.704474 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.704531 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.704541 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.704558 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.704571 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:45Z","lastTransitionTime":"2025-10-04T03:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.807472 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.807596 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.807612 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.807633 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.807644 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:45Z","lastTransitionTime":"2025-10-04T03:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.857844 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:11:45 crc kubenswrapper[4742]: E1004 03:11:45.858058 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.910414 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.910731 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.910743 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.910764 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:45 crc kubenswrapper[4742]: I1004 03:11:45.910775 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:45Z","lastTransitionTime":"2025-10-04T03:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.014338 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.014389 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.014399 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.014421 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.014433 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:46Z","lastTransitionTime":"2025-10-04T03:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.117716 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.117771 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.117784 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.117805 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.117817 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:46Z","lastTransitionTime":"2025-10-04T03:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.220442 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.220529 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.220546 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.220602 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.220619 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:46Z","lastTransitionTime":"2025-10-04T03:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.271800 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nc8m7_46de52ee-226c-4713-bd7b-961f8f43f635/kube-multus/0.log" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.271861 4742 generic.go:334] "Generic (PLEG): container finished" podID="46de52ee-226c-4713-bd7b-961f8f43f635" containerID="7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c" exitCode=1 Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.271907 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nc8m7" event={"ID":"46de52ee-226c-4713-bd7b-961f8f43f635","Type":"ContainerDied","Data":"7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c"} Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.272422 4742 scope.go:117] "RemoveContainer" containerID="7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.288227 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://261e8ace3d214aec7b737ea538ac0996cc11a5dae65eacc33dd3a3ea65ed3558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.308525 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd1d6092975dd4703ff4930e2eafaf464f221510fddeda0f977a73c548c305eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.323163 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cwm6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a5aee9-b917-473a-9196-5cde6553c72a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6cefd4be7726c02ec9a7b2db28ce61c5dd11b708a131e05e96595b51719209b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrdmr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cwm6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.326566 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.326643 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.326655 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.326678 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.326693 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:46Z","lastTransitionTime":"2025-10-04T03:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.339599 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cd46x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0302950c-deb3-420a-b798-c99459e62b3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fls9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fls9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:12Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cd46x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.355167 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ec6cadd-fdb4-43c5-a898-f1bde3e9898d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf178a15abcb94eda5033b74ef132d9e204b33f4112152edf172e61999fc2c9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac618416ef312d2c051d88c0348fea230d73d3c8d6f4b1830b4b584ceae645c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2ede8f31552e98a3087393a27a4bab2c56f116829fef768b9c1dda8d6630142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3ee868ef16effadb77812f1673237d65d10493bd29dfcef5c7fe53da640d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce3ee868ef16effadb77812f1673237d65d10493bd29dfcef5c7fe53da640d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.372740 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.388749 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11b178c3523269e8b4c09a1ad6f86eaee1d8f19f5056840dfa24466248889a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.402758 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.419162 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.429851 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.429891 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.429904 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.429924 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.429937 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:46Z","lastTransitionTime":"2025-10-04T03:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.437969 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.451784 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.464972 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc8m7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46de52ee-226c-4713-bd7b-961f8f43f635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:11:45Z\\\",\\\"message\\\":\\\"2025-10-04T03:11:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a40fc70a-b392-4173-9861-b7ae2db0575b\\\\n2025-10-04T03:11:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a40fc70a-b392-4173-9861-b7ae2db0575b to /host/opt/cni/bin/\\\\n2025-10-04T03:11:00Z [verbose] multus-daemon started\\\\n2025-10-04T03:11:00Z [verbose] Readiness Indicator file check\\\\n2025-10-04T03:11:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pj2t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc8m7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.477978 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2694d67-7fb5-446f-a590-6113e9131921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b29154af28bbb704eeb5fde7a5c44db624f1cb35d79942c62d5ca9a03e697b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1070bde4d35430f4c712d63ede34a428ec20a277ce0409dd3e69dead27c3f98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwzp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.495648 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.509428 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.524662 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.541108 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.541181 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.541206 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.541239 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.541258 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:46Z","lastTransitionTime":"2025-10-04T03:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.563257 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12a2784adf6efd5697d0afb2d3a66ea54c2b234d802f0065d6839be9d91fa920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12a2784adf6efd5697d0afb2d3a66ea54c2b234d802f0065d6839be9d91fa920\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:11:21Z\\\",\\\"message\\\":\\\"0-73dcf8201ba5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 03:11:21.773108 6407 services_controller.go:451] Built service openshift-kube-storage-version-migrator-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1004 03:11:21.773205 6407 services_controller.go:452] Built service openshift-kube-storage-version-migrator-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1004 03:11:21.773214 6407 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/mac\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9qrtl_openshift-ovn-kubernetes(0cfc6e64-ba08-4c44-a18e-3b93ab792a6a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qrtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.644519 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.644593 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.644616 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.644643 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.644664 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:46Z","lastTransitionTime":"2025-10-04T03:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.747342 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.747411 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.747425 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.747445 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.747456 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:46Z","lastTransitionTime":"2025-10-04T03:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.850308 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.850353 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.850364 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.850382 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.850392 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:46Z","lastTransitionTime":"2025-10-04T03:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.857562 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.857595 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:46 crc kubenswrapper[4742]: E1004 03:11:46.857705 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.857725 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:11:46 crc kubenswrapper[4742]: E1004 03:11:46.857875 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:11:46 crc kubenswrapper[4742]: E1004 03:11:46.858123 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.952959 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.953000 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.953010 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.953026 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:46 crc kubenswrapper[4742]: I1004 03:11:46.953036 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:46Z","lastTransitionTime":"2025-10-04T03:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.055692 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.055745 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.055756 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.055773 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.055784 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:47Z","lastTransitionTime":"2025-10-04T03:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.158130 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.158179 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.158191 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.158210 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.158222 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:47Z","lastTransitionTime":"2025-10-04T03:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.261167 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.261219 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.261228 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.261249 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.261260 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:47Z","lastTransitionTime":"2025-10-04T03:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.277366 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nc8m7_46de52ee-226c-4713-bd7b-961f8f43f635/kube-multus/0.log" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.277428 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nc8m7" event={"ID":"46de52ee-226c-4713-bd7b-961f8f43f635","Type":"ContainerStarted","Data":"bfe5cd4a2153e5ae33beedc1cfdee7c39f57d2dc41a3d5e7d442df75d3c069db"} Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.305153 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12a2784adf6efd5697d0afb2d3a66ea54c2b234d802f0065d6839be9d91fa920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12a2784adf6efd5697d0afb2d3a66ea54c2b234d802f0065d6839be9d91fa920\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:11:21Z\\\",\\\"message\\\":\\\"0-73dcf8201ba5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 03:11:21.773108 6407 services_controller.go:451] Built service openshift-kube-storage-version-migrator-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1004 03:11:21.773205 6407 services_controller.go:452] Built service openshift-kube-storage-version-migrator-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1004 03:11:21.773214 6407 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/mac\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9qrtl_openshift-ovn-kubernetes(0cfc6e64-ba08-4c44-a18e-3b93ab792a6a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qrtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.318795 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.331638 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.343431 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.354608 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cd46x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0302950c-deb3-420a-b798-c99459e62b3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fls9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fls9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:12Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cd46x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.364095 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.364167 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.364187 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.364218 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.364238 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:47Z","lastTransitionTime":"2025-10-04T03:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.365021 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://261e8ace3d214aec7b737ea538ac0996cc11a5dae65eacc33dd3a3ea65ed3558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.380401 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd1d6092975dd4703ff4930e2eafaf464f221510fddeda0f977a73c548c305eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.391650 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cwm6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a5aee9-b917-473a-9196-5cde6553c72a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6cefd4be7726c02ec9a7b2db28ce61c5dd11b708a131e05e96595b51719209b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrdmr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cwm6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.403594 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ec6cadd-fdb4-43c5-a898-f1bde3e9898d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf178a15abcb94eda5033b74ef132d9e204b33f4112152edf172e61999fc2c9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac618416ef312d2c051d88c0348fea230d73d3c8d6f4b1830b4b584ceae645c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2ede8f31552e98a3087393a27a4bab2c56f116829fef768b9c1dda8d6630142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3ee868ef16effadb77812f1673237d65d10493bd29dfcef5c7fe53da640d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce3ee868ef16effadb77812f1673237d65d10493bd29dfcef5c7fe53da640d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.418218 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.427826 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11b178c3523269e8b4c09a1ad6f86eaee1d8f19f5056840dfa24466248889a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.442401 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc8m7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46de52ee-226c-4713-bd7b-961f8f43f635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe5cd4a2153e5ae33beedc1cfdee7c39f57d2dc41a3d5e7d442df75d3c069db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:11:45Z\\\",\\\"message\\\":\\\"2025-10-04T03:11:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a40fc70a-b392-4173-9861-b7ae2db0575b\\\\n2025-10-04T03:11:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a40fc70a-b392-4173-9861-b7ae2db0575b to /host/opt/cni/bin/\\\\n2025-10-04T03:11:00Z [verbose] multus-daemon started\\\\n2025-10-04T03:11:00Z [verbose] Readiness Indicator file check\\\\n2025-10-04T03:11:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pj2t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc8m7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.455329 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2694d67-7fb5-446f-a590-6113e9131921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b29154af28bbb704eeb5fde7a5c44db624f1cb35d79942c62d5ca9a03e697b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1070bde4d35430f4c712d63ede34a428ec20a277ce0409dd3e69dead27c3f98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwzp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.466500 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.466577 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.466604 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.466628 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.466643 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:47Z","lastTransitionTime":"2025-10-04T03:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.471246 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.484425 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.495353 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.506499 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.569531 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.569571 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.569580 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.569599 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.569608 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:47Z","lastTransitionTime":"2025-10-04T03:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.672294 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.672349 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.672363 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.672380 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.672391 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:47Z","lastTransitionTime":"2025-10-04T03:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.775268 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.775328 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.775336 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.775354 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.775363 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:47Z","lastTransitionTime":"2025-10-04T03:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.857771 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:11:47 crc kubenswrapper[4742]: E1004 03:11:47.857956 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.877740 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.877791 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.877803 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.877823 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.877837 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:47Z","lastTransitionTime":"2025-10-04T03:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.980170 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.980238 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.980252 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.980301 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:47 crc kubenswrapper[4742]: I1004 03:11:47.980320 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:47Z","lastTransitionTime":"2025-10-04T03:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.083639 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.083730 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.084388 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.084439 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.084459 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:48Z","lastTransitionTime":"2025-10-04T03:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.187953 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.188002 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.188015 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.188034 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.188046 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:48Z","lastTransitionTime":"2025-10-04T03:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.290798 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.290857 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.290876 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.290905 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.290924 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:48Z","lastTransitionTime":"2025-10-04T03:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.394404 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.394513 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.394540 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.394592 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.394612 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:48Z","lastTransitionTime":"2025-10-04T03:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.497228 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.497314 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.497329 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.497351 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.497366 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:48Z","lastTransitionTime":"2025-10-04T03:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.600485 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.600538 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.600547 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.600568 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.600578 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:48Z","lastTransitionTime":"2025-10-04T03:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.704314 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.704421 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.704443 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.704468 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.704487 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:48Z","lastTransitionTime":"2025-10-04T03:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.807812 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.807856 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.807866 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.807886 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.807896 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:48Z","lastTransitionTime":"2025-10-04T03:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.859512 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.859512 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.859627 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:11:48 crc kubenswrapper[4742]: E1004 03:11:48.860084 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:11:48 crc kubenswrapper[4742]: E1004 03:11:48.860193 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.860396 4742 scope.go:117] "RemoveContainer" containerID="12a2784adf6efd5697d0afb2d3a66ea54c2b234d802f0065d6839be9d91fa920" Oct 04 03:11:48 crc kubenswrapper[4742]: E1004 03:11:48.860476 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.911051 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.911173 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.911231 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.911264 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:48 crc kubenswrapper[4742]: I1004 03:11:48.911335 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:48Z","lastTransitionTime":"2025-10-04T03:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.014709 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.014749 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.014762 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.014781 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.014795 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:49Z","lastTransitionTime":"2025-10-04T03:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.117703 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.117909 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.118042 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.118107 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.118124 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:49Z","lastTransitionTime":"2025-10-04T03:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.221611 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.221655 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.221666 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.221684 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.221696 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:49Z","lastTransitionTime":"2025-10-04T03:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.287040 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qrtl_0cfc6e64-ba08-4c44-a18e-3b93ab792a6a/ovnkube-controller/2.log" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.290748 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" event={"ID":"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a","Type":"ContainerStarted","Data":"9b37c3bfb97a4dc710d7664dfa0bc1cd0093419f21a8ca483ac1f38e695d7271"} Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.291618 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.319307 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.324162 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.324197 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.324210 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.324232 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.324247 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:49Z","lastTransitionTime":"2025-10-04T03:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.338399 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.361793 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.386064 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.401590 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc8m7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46de52ee-226c-4713-bd7b-961f8f43f635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe5cd4a2153e5ae33beedc1cfdee7c39f57d2dc41a3d5e7d442df75d3c069db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:11:45Z\\\",\\\"message\\\":\\\"2025-10-04T03:11:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a40fc70a-b392-4173-9861-b7ae2db0575b\\\\n2025-10-04T03:11:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a40fc70a-b392-4173-9861-b7ae2db0575b to /host/opt/cni/bin/\\\\n2025-10-04T03:11:00Z [verbose] multus-daemon started\\\\n2025-10-04T03:11:00Z [verbose] Readiness Indicator file check\\\\n2025-10-04T03:11:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pj2t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc8m7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.414546 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2694d67-7fb5-446f-a590-6113e9131921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b29154af28bbb704eeb5fde7a5c44db624f1cb35d79942c62d5ca9a03e697b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1070bde4d35430f4c712d63ede34a428ec20a277ce0409dd3e69dead27c3f98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwzp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.427221 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.427312 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.427330 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.427350 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.427365 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:49Z","lastTransitionTime":"2025-10-04T03:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.429684 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.447562 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.460716 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.491198 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b37c3bfb97a4dc710d7664dfa0bc1cd0093419f21a8ca483ac1f38e695d7271\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12a2784adf6efd5697d0afb2d3a66ea54c2b234d802f0065d6839be9d91fa920\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:11:21Z\\\",\\\"message\\\":\\\"0-73dcf8201ba5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 03:11:21.773108 6407 services_controller.go:451] Built service openshift-kube-storage-version-migrator-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1004 03:11:21.773205 6407 services_controller.go:452] Built service openshift-kube-storage-version-migrator-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1004 03:11:21.773214 6407 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/mac\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qrtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.505638 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://261e8ace3d214aec7b737ea538ac0996cc11a5dae65eacc33dd3a3ea65ed3558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.521704 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd1d6092975dd4703ff4930e2eafaf464f221510fddeda0f977a73c548c305eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.529999 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.530051 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.530062 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.530132 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.530144 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:49Z","lastTransitionTime":"2025-10-04T03:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.533699 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cwm6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a5aee9-b917-473a-9196-5cde6553c72a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6cefd4be7726c02ec9a7b2db28ce61c5dd11b708a131e05e96595b51719209b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrdmr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cwm6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.548302 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cd46x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0302950c-deb3-420a-b798-c99459e62b3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fls9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fls9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:12Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cd46x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.568795 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ec6cadd-fdb4-43c5-a898-f1bde3e9898d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf178a15abcb94eda5033b74ef132d9e204b33f4112152edf172e61999fc2c9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac618416ef312d2c051d88c0348fea230d73d3c8d6f4b1830b4b584ceae645c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2ede8f31552e98a3087393a27a4bab2c56f116829fef768b9c1dda8d6630142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3ee868ef16effadb77812f1673237d65d10493bd29dfcef5c7fe53da640d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce3ee868ef16effadb77812f1673237d65d10493bd29dfcef5c7fe53da640d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.585128 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.595841 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11b178c3523269e8b4c09a1ad6f86eaee1d8f19f5056840dfa24466248889a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.632720 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.632773 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.632785 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.632801 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.632813 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:49Z","lastTransitionTime":"2025-10-04T03:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.735501 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.735568 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.735579 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.735597 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.735609 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:49Z","lastTransitionTime":"2025-10-04T03:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.838452 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.838527 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.838537 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.838553 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.838567 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:49Z","lastTransitionTime":"2025-10-04T03:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.857597 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:11:49 crc kubenswrapper[4742]: E1004 03:11:49.857719 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.940958 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.941007 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.941021 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.941039 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:49 crc kubenswrapper[4742]: I1004 03:11:49.941051 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:49Z","lastTransitionTime":"2025-10-04T03:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.043399 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.043439 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.043448 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.043466 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.043476 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:50Z","lastTransitionTime":"2025-10-04T03:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.145713 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.145752 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.145761 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.145781 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.145791 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:50Z","lastTransitionTime":"2025-10-04T03:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.248354 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.248402 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.248416 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.248430 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.248443 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:50Z","lastTransitionTime":"2025-10-04T03:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.296174 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qrtl_0cfc6e64-ba08-4c44-a18e-3b93ab792a6a/ovnkube-controller/3.log" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.296841 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qrtl_0cfc6e64-ba08-4c44-a18e-3b93ab792a6a/ovnkube-controller/2.log" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.299408 4742 generic.go:334] "Generic (PLEG): container finished" podID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerID="9b37c3bfb97a4dc710d7664dfa0bc1cd0093419f21a8ca483ac1f38e695d7271" exitCode=1 Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.299460 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" event={"ID":"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a","Type":"ContainerDied","Data":"9b37c3bfb97a4dc710d7664dfa0bc1cd0093419f21a8ca483ac1f38e695d7271"} Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.299509 4742 scope.go:117] "RemoveContainer" containerID="12a2784adf6efd5697d0afb2d3a66ea54c2b234d802f0065d6839be9d91fa920" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.300499 4742 scope.go:117] "RemoveContainer" containerID="9b37c3bfb97a4dc710d7664dfa0bc1cd0093419f21a8ca483ac1f38e695d7271" Oct 04 03:11:50 crc kubenswrapper[4742]: E1004 03:11:50.301551 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9qrtl_openshift-ovn-kubernetes(0cfc6e64-ba08-4c44-a18e-3b93ab792a6a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.313745 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cd46x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0302950c-deb3-420a-b798-c99459e62b3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fls9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fls9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:12Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cd46x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.327036 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://261e8ace3d214aec7b737ea538ac0996cc11a5dae65eacc33dd3a3ea65ed3558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.340353 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd1d6092975dd4703ff4930e2eafaf464f221510fddeda0f977a73c548c305eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.351347 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.351422 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.351443 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.351472 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.351492 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:50Z","lastTransitionTime":"2025-10-04T03:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.351928 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cwm6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a5aee9-b917-473a-9196-5cde6553c72a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6cefd4be7726c02ec9a7b2db28ce61c5dd11b708a131e05e96595b51719209b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrdmr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cwm6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.364377 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ec6cadd-fdb4-43c5-a898-f1bde3e9898d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf178a15abcb94eda5033b74ef132d9e204b33f4112152edf172e61999fc2c9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac618416ef312d2c051d88c0348fea230d73d3c8d6f4b1830b4b584ceae645c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2ede8f31552e98a3087393a27a4bab2c56f116829fef768b9c1dda8d6630142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3ee868ef16effadb77812f1673237d65d10493bd29dfcef5c7fe53da640d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce3ee868ef16effadb77812f1673237d65d10493bd29dfcef5c7fe53da640d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.375747 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.383849 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11b178c3523269e8b4c09a1ad6f86eaee1d8f19f5056840dfa24466248889a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.394376 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc8m7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46de52ee-226c-4713-bd7b-961f8f43f635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe5cd4a2153e5ae33beedc1cfdee7c39f57d2dc41a3d5e7d442df75d3c069db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:11:45Z\\\",\\\"message\\\":\\\"2025-10-04T03:11:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a40fc70a-b392-4173-9861-b7ae2db0575b\\\\n2025-10-04T03:11:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a40fc70a-b392-4173-9861-b7ae2db0575b to /host/opt/cni/bin/\\\\n2025-10-04T03:11:00Z [verbose] multus-daemon started\\\\n2025-10-04T03:11:00Z [verbose] Readiness Indicator file check\\\\n2025-10-04T03:11:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pj2t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc8m7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.403077 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2694d67-7fb5-446f-a590-6113e9131921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b29154af28bbb704eeb5fde7a5c44db624f1cb35d79942c62d5ca9a03e697b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1070bde4d35430f4c712d63ede34a428ec20a277ce0409dd3e69dead27c3f98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwzp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.413224 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.425798 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.436696 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.446408 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.454535 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.454566 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.454577 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.454594 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.454622 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:50Z","lastTransitionTime":"2025-10-04T03:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.464497 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b37c3bfb97a4dc710d7664dfa0bc1cd0093419f21a8ca483ac1f38e695d7271\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12a2784adf6efd5697d0afb2d3a66ea54c2b234d802f0065d6839be9d91fa920\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:11:21Z\\\",\\\"message\\\":\\\"0-73dcf8201ba5}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 03:11:21.773108 6407 services_controller.go:451] Built service openshift-kube-storage-version-migrator-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1004 03:11:21.773205 6407 services_controller.go:452] Built service openshift-kube-storage-version-migrator-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1004 03:11:21.773214 6407 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-machine-webhook]} name:Service_openshift-machine-api/mac\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b37c3bfb97a4dc710d7664dfa0bc1cd0093419f21a8ca483ac1f38e695d7271\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:11:49Z\\\",\\\"message\\\":\\\"_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1004 03:11:49.775768 6758 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-nc8m7\\\\nI1004 03:11:49.775649 6758 services_controller.go:434] Service openshift-machine-api/control-plane-machine-set-operator retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{control-plane-machine-set-operator openshift-machine-api ffd0ef27-d28d-43cc-90c8-0e8843e4c04c 4409 0 2025-02-23 05:12:21 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:control-plane-machine-set-operator] map[capability.openshift.io/name:MachineAPI exclude.release.openshift.io/internal-openshift-hosted:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:control-plane-machine-set-operator-tls service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0077282e7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:9443,TargetPort:{1 0 https},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{k8s-app: control-plane-machine-set-operator,},ClusterIP:10.21\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qrtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.480049 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.492373 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.502799 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.556696 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.556748 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.556763 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.556783 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.556796 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:50Z","lastTransitionTime":"2025-10-04T03:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.659781 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.659823 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.659835 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.659852 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.659866 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:50Z","lastTransitionTime":"2025-10-04T03:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.762312 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.762344 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.762355 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.762374 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.762386 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:50Z","lastTransitionTime":"2025-10-04T03:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.857886 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.857947 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:50 crc kubenswrapper[4742]: E1004 03:11:50.858042 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.858054 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:11:50 crc kubenswrapper[4742]: E1004 03:11:50.858186 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:11:50 crc kubenswrapper[4742]: E1004 03:11:50.858264 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.864940 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.864985 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.864998 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.865018 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.865031 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:50Z","lastTransitionTime":"2025-10-04T03:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.969860 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.969908 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.969938 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.969969 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:50 crc kubenswrapper[4742]: I1004 03:11:50.969991 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:50Z","lastTransitionTime":"2025-10-04T03:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.072787 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.072842 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.072855 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.072876 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.072891 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:51Z","lastTransitionTime":"2025-10-04T03:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.175791 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.175843 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.175855 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.175875 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.175887 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:51Z","lastTransitionTime":"2025-10-04T03:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.278125 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.278157 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.278167 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.278182 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.278192 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:51Z","lastTransitionTime":"2025-10-04T03:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.304857 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qrtl_0cfc6e64-ba08-4c44-a18e-3b93ab792a6a/ovnkube-controller/3.log" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.308399 4742 scope.go:117] "RemoveContainer" containerID="9b37c3bfb97a4dc710d7664dfa0bc1cd0093419f21a8ca483ac1f38e695d7271" Oct 04 03:11:51 crc kubenswrapper[4742]: E1004 03:11:51.308574 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9qrtl_openshift-ovn-kubernetes(0cfc6e64-ba08-4c44-a18e-3b93ab792a6a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.319619 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://261e8ace3d214aec7b737ea538ac0996cc11a5dae65eacc33dd3a3ea65ed3558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.336149 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd1d6092975dd4703ff4930e2eafaf464f221510fddeda0f977a73c548c305eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.346939 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cwm6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a5aee9-b917-473a-9196-5cde6553c72a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6cefd4be7726c02ec9a7b2db28ce61c5dd11b708a131e05e96595b51719209b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrdmr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cwm6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.357147 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cd46x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0302950c-deb3-420a-b798-c99459e62b3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fls9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fls9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:12Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cd46x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.368480 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ec6cadd-fdb4-43c5-a898-f1bde3e9898d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf178a15abcb94eda5033b74ef132d9e204b33f4112152edf172e61999fc2c9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac618416ef312d2c051d88c0348fea230d73d3c8d6f4b1830b4b584ceae645c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2ede8f31552e98a3087393a27a4bab2c56f116829fef768b9c1dda8d6630142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3ee868ef16effadb77812f1673237d65d10493bd29dfcef5c7fe53da640d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce3ee868ef16effadb77812f1673237d65d10493bd29dfcef5c7fe53da640d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.380804 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.380912 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.380923 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.380944 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.380958 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:51Z","lastTransitionTime":"2025-10-04T03:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.387527 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.401933 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11b178c3523269e8b4c09a1ad6f86eaee1d8f19f5056840dfa24466248889a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.418677 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.433001 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.445422 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.456459 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.468413 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc8m7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46de52ee-226c-4713-bd7b-961f8f43f635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe5cd4a2153e5ae33beedc1cfdee7c39f57d2dc41a3d5e7d442df75d3c069db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:11:45Z\\\",\\\"message\\\":\\\"2025-10-04T03:11:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a40fc70a-b392-4173-9861-b7ae2db0575b\\\\n2025-10-04T03:11:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a40fc70a-b392-4173-9861-b7ae2db0575b to /host/opt/cni/bin/\\\\n2025-10-04T03:11:00Z [verbose] multus-daemon started\\\\n2025-10-04T03:11:00Z [verbose] Readiness Indicator file check\\\\n2025-10-04T03:11:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pj2t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc8m7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.481196 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2694d67-7fb5-446f-a590-6113e9131921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b29154af28bbb704eeb5fde7a5c44db624f1cb35d79942c62d5ca9a03e697b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1070bde4d35430f4c712d63ede34a428ec20a277ce0409dd3e69dead27c3f98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwzp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.482776 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.482884 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.482952 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.483034 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.483140 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:51Z","lastTransitionTime":"2025-10-04T03:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.494428 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.508467 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.521263 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.537970 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b37c3bfb97a4dc710d7664dfa0bc1cd0093419f21a8ca483ac1f38e695d7271\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b37c3bfb97a4dc710d7664dfa0bc1cd0093419f21a8ca483ac1f38e695d7271\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:11:49Z\\\",\\\"message\\\":\\\"_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1004 03:11:49.775768 6758 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-nc8m7\\\\nI1004 03:11:49.775649 6758 services_controller.go:434] Service openshift-machine-api/control-plane-machine-set-operator retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{control-plane-machine-set-operator openshift-machine-api ffd0ef27-d28d-43cc-90c8-0e8843e4c04c 4409 0 2025-02-23 05:12:21 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:control-plane-machine-set-operator] map[capability.openshift.io/name:MachineAPI exclude.release.openshift.io/internal-openshift-hosted:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:control-plane-machine-set-operator-tls service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0077282e7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:9443,TargetPort:{1 0 https},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{k8s-app: control-plane-machine-set-operator,},ClusterIP:10.21\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9qrtl_openshift-ovn-kubernetes(0cfc6e64-ba08-4c44-a18e-3b93ab792a6a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qrtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.585950 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.586087 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.586164 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.586233 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.586331 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:51Z","lastTransitionTime":"2025-10-04T03:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.688083 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.688148 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.688168 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.688195 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.688220 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:51Z","lastTransitionTime":"2025-10-04T03:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.790302 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.790376 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.790398 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.790425 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.790450 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:51Z","lastTransitionTime":"2025-10-04T03:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.849430 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.849486 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.849504 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.849524 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.849535 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:51Z","lastTransitionTime":"2025-10-04T03:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.857575 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:11:51 crc kubenswrapper[4742]: E1004 03:11:51.857776 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:11:51 crc kubenswrapper[4742]: E1004 03:11:51.862353 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.865747 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.865778 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.865787 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.865803 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.865813 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:51Z","lastTransitionTime":"2025-10-04T03:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:51 crc kubenswrapper[4742]: E1004 03:11:51.878735 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.881971 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.882010 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.882024 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.882042 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.882054 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:51Z","lastTransitionTime":"2025-10-04T03:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:51 crc kubenswrapper[4742]: E1004 03:11:51.894422 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.899043 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.899165 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.899231 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.899324 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.899408 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:51Z","lastTransitionTime":"2025-10-04T03:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:51 crc kubenswrapper[4742]: E1004 03:11:51.910365 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.914192 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.914240 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.914251 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.914287 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.914300 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:51Z","lastTransitionTime":"2025-10-04T03:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:51 crc kubenswrapper[4742]: E1004 03:11:51.924905 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:51 crc kubenswrapper[4742]: E1004 03:11:51.925023 4742 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.926476 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.926501 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.926511 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.926526 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:51 crc kubenswrapper[4742]: I1004 03:11:51.926535 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:51Z","lastTransitionTime":"2025-10-04T03:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.029097 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.029152 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.029166 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.029190 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.029203 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:52Z","lastTransitionTime":"2025-10-04T03:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.136336 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.136433 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.136444 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.136464 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.136475 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:52Z","lastTransitionTime":"2025-10-04T03:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.239283 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.239335 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.239349 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.239368 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.239381 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:52Z","lastTransitionTime":"2025-10-04T03:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.342245 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.342306 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.342317 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.342331 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.342341 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:52Z","lastTransitionTime":"2025-10-04T03:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.444737 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.444799 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.444814 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.444835 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.444849 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:52Z","lastTransitionTime":"2025-10-04T03:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.547564 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.547616 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.547626 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.547645 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.547659 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:52Z","lastTransitionTime":"2025-10-04T03:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.650145 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.650204 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.650216 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.650233 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.650245 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:52Z","lastTransitionTime":"2025-10-04T03:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.752812 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.752854 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.752864 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.752879 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.752889 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:52Z","lastTransitionTime":"2025-10-04T03:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.855148 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.855194 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.855204 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.855240 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.855251 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:52Z","lastTransitionTime":"2025-10-04T03:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.857888 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.857948 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.857951 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:52 crc kubenswrapper[4742]: E1004 03:11:52.858049 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:11:52 crc kubenswrapper[4742]: E1004 03:11:52.858201 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:11:52 crc kubenswrapper[4742]: E1004 03:11:52.858229 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.958928 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.959300 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.959436 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.959778 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:52 crc kubenswrapper[4742]: I1004 03:11:52.960064 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:52Z","lastTransitionTime":"2025-10-04T03:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.063179 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.063672 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.063839 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.064001 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.064141 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:53Z","lastTransitionTime":"2025-10-04T03:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.167621 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.167671 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.167685 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.167704 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.167717 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:53Z","lastTransitionTime":"2025-10-04T03:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.271000 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.271051 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.271062 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.271083 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.271096 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:53Z","lastTransitionTime":"2025-10-04T03:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.374761 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.374820 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.374838 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.374866 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.374883 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:53Z","lastTransitionTime":"2025-10-04T03:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.477163 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.477260 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.477329 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.477378 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.477406 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:53Z","lastTransitionTime":"2025-10-04T03:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.580334 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.580374 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.580382 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.580397 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.580406 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:53Z","lastTransitionTime":"2025-10-04T03:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.683413 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.683463 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.683474 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.683678 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.683693 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:53Z","lastTransitionTime":"2025-10-04T03:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.786140 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.786176 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.786186 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.786202 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.786212 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:53Z","lastTransitionTime":"2025-10-04T03:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.857601 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:11:53 crc kubenswrapper[4742]: E1004 03:11:53.857753 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.869769 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.888534 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.888568 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.888578 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.888595 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.888604 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:53Z","lastTransitionTime":"2025-10-04T03:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.991774 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.991820 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.991836 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.991861 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:53 crc kubenswrapper[4742]: I1004 03:11:53.991878 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:53Z","lastTransitionTime":"2025-10-04T03:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.095123 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.095544 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.095553 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.095576 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.095588 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:54Z","lastTransitionTime":"2025-10-04T03:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.198104 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.198204 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.198230 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.198300 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.198324 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:54Z","lastTransitionTime":"2025-10-04T03:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.301638 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.301700 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.301713 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.301733 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.301745 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:54Z","lastTransitionTime":"2025-10-04T03:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.404607 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.404656 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.404669 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.404692 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.404705 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:54Z","lastTransitionTime":"2025-10-04T03:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.507789 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.507843 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.507858 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.507877 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.507891 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:54Z","lastTransitionTime":"2025-10-04T03:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.610926 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.611015 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.611028 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.611049 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.611061 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:54Z","lastTransitionTime":"2025-10-04T03:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.713586 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.713656 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.713666 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.713685 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.713697 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:54Z","lastTransitionTime":"2025-10-04T03:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.816365 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.816434 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.816444 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.816462 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.816472 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:54Z","lastTransitionTime":"2025-10-04T03:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.857933 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.858072 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:11:54 crc kubenswrapper[4742]: E1004 03:11:54.858183 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.858209 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:54 crc kubenswrapper[4742]: E1004 03:11:54.858378 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:11:54 crc kubenswrapper[4742]: E1004 03:11:54.858543 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.870151 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.870776 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11b178c3523269e8b4c09a1ad6f86eaee1d8f19f5056840dfa24466248889a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.884552 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ec6cadd-fdb4-43c5-a898-f1bde3e9898d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf178a15abcb94eda5033b74ef132d9e204b33f4112152edf172e61999fc2c9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac618416ef312d2c051d88c0348fea230d73d3c8d6f4b1830b4b584ceae645c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2ede8f31552e98a3087393a27a4bab2c56f116829fef768b9c1dda8d6630142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3ee868ef16effadb77812f1673237d65d10493bd29dfcef5c7fe53da640d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce3ee868ef16effadb77812f1673237d65d10493bd29dfcef5c7fe53da640d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.899151 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.912756 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.919874 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.919936 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.919974 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.920003 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.920037 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:54Z","lastTransitionTime":"2025-10-04T03:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.924768 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.938733 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc8m7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46de52ee-226c-4713-bd7b-961f8f43f635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe5cd4a2153e5ae33beedc1cfdee7c39f57d2dc41a3d5e7d442df75d3c069db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:11:45Z\\\",\\\"message\\\":\\\"2025-10-04T03:11:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a40fc70a-b392-4173-9861-b7ae2db0575b\\\\n2025-10-04T03:11:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a40fc70a-b392-4173-9861-b7ae2db0575b to /host/opt/cni/bin/\\\\n2025-10-04T03:11:00Z [verbose] multus-daemon started\\\\n2025-10-04T03:11:00Z [verbose] Readiness Indicator file check\\\\n2025-10-04T03:11:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pj2t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc8m7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.956022 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2694d67-7fb5-446f-a590-6113e9131921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b29154af28bbb704eeb5fde7a5c44db624f1cb35d79942c62d5ca9a03e697b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1070bde4d35430f4c712d63ede34a428ec20a277ce0409dd3e69dead27c3f98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwzp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.973324 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:54 crc kubenswrapper[4742]: I1004 03:11:54.990034 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.007347 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:55Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.024341 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.024425 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.024449 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.024478 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.024498 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:55Z","lastTransitionTime":"2025-10-04T03:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.026530 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:55Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.048217 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b37c3bfb97a4dc710d7664dfa0bc1cd0093419f21a8ca483ac1f38e695d7271\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b37c3bfb97a4dc710d7664dfa0bc1cd0093419f21a8ca483ac1f38e695d7271\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:11:49Z\\\",\\\"message\\\":\\\"_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1004 03:11:49.775768 6758 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-nc8m7\\\\nI1004 03:11:49.775649 6758 services_controller.go:434] Service openshift-machine-api/control-plane-machine-set-operator retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{control-plane-machine-set-operator openshift-machine-api ffd0ef27-d28d-43cc-90c8-0e8843e4c04c 4409 0 2025-02-23 05:12:21 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:control-plane-machine-set-operator] map[capability.openshift.io/name:MachineAPI exclude.release.openshift.io/internal-openshift-hosted:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:control-plane-machine-set-operator-tls service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0077282e7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:9443,TargetPort:{1 0 https},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{k8s-app: control-plane-machine-set-operator,},ClusterIP:10.21\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9qrtl_openshift-ovn-kubernetes(0cfc6e64-ba08-4c44-a18e-3b93ab792a6a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qrtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:55Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.068737 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:55Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.091795 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd1d6092975dd4703ff4930e2eafaf464f221510fddeda0f977a73c548c305eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:55Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.114544 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cwm6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a5aee9-b917-473a-9196-5cde6553c72a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6cefd4be7726c02ec9a7b2db28ce61c5dd11b708a131e05e96595b51719209b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrdmr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cwm6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:55Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.128309 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.128410 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.128435 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.128566 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.128660 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:55Z","lastTransitionTime":"2025-10-04T03:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.147571 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cd46x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0302950c-deb3-420a-b798-c99459e62b3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fls9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fls9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:12Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cd46x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:55Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.190362 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91bc9f85-8f09-4d65-a81f-c8e50b0fbdb1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://830fe4584483d987ccd84235d0d565579604c385ffc648b6a61868a7369a19c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65ba85e90248976f60214ebcda8e19348e9c373aa509ca521cb8e313acec84c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01061b91a5279d579507bc180cc59378d3729dc8d327f90a37f3e89331d6a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f53f64a2c0a1b813ae42baac78f5e639dfc576212d66b48480b0d3563120e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://467455dfc35412e2c98ab78be938378877791ef8e6acd843795dd5e600495934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48906a72bb35674544fa9d074de2f582da4cbad41e0472f6bea749cd6a7d860c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48906a72bb35674544fa9d074de2f582da4cbad41e0472f6bea749cd6a7d860c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9bcffac1019b9253a814cd8b0951819a40309057f15c54bf6ff4ee0fc491e61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9bcffac1019b9253a814cd8b0951819a40309057f15c54bf6ff4ee0fc491e61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0fd66b11b8abc5ab89eb8355cba078dc46e2be3ee0f59afc37cdcee8e80115b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd66b11b8abc5ab89eb8355cba078dc46e2be3ee0f59afc37cdcee8e80115b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:55Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.204770 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://261e8ace3d214aec7b737ea538ac0996cc11a5dae65eacc33dd3a3ea65ed3558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:11:55Z is after 2025-08-24T17:21:41Z" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.231521 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.231574 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.231587 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.231606 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.231858 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:55Z","lastTransitionTime":"2025-10-04T03:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.335328 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.335384 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.335394 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.335412 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.335424 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:55Z","lastTransitionTime":"2025-10-04T03:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.439248 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.439331 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.439345 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.439367 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.439381 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:55Z","lastTransitionTime":"2025-10-04T03:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.542529 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.542766 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.542778 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.542799 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.542812 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:55Z","lastTransitionTime":"2025-10-04T03:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.646571 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.646630 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.646647 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.646671 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.646688 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:55Z","lastTransitionTime":"2025-10-04T03:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.749498 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.749549 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.749559 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.749579 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.749593 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:55Z","lastTransitionTime":"2025-10-04T03:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.853218 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.853341 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.853373 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.853402 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.853420 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:55Z","lastTransitionTime":"2025-10-04T03:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.857827 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:11:55 crc kubenswrapper[4742]: E1004 03:11:55.858028 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.957123 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.957184 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.957201 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.957229 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:55 crc kubenswrapper[4742]: I1004 03:11:55.957251 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:55Z","lastTransitionTime":"2025-10-04T03:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.060586 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.060679 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.060703 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.060739 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.060763 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:56Z","lastTransitionTime":"2025-10-04T03:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.164365 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.164432 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.164441 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.164459 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.164469 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:56Z","lastTransitionTime":"2025-10-04T03:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.268562 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.268644 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.268662 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.268695 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.268713 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:56Z","lastTransitionTime":"2025-10-04T03:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.371598 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.371651 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.371661 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.371679 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.371691 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:56Z","lastTransitionTime":"2025-10-04T03:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.474401 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.474462 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.474475 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.474498 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.474517 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:56Z","lastTransitionTime":"2025-10-04T03:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.576625 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.576660 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.576686 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.576702 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.576712 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:56Z","lastTransitionTime":"2025-10-04T03:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.680104 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.680174 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.680188 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.680210 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.680224 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:56Z","lastTransitionTime":"2025-10-04T03:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.783631 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.783720 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.783771 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.783798 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.783816 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:56Z","lastTransitionTime":"2025-10-04T03:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.858567 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.858622 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.858765 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:11:56 crc kubenswrapper[4742]: E1004 03:11:56.858961 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:11:56 crc kubenswrapper[4742]: E1004 03:11:56.859128 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:11:56 crc kubenswrapper[4742]: E1004 03:11:56.859335 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.887589 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.887663 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.887682 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.887711 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.887729 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:56Z","lastTransitionTime":"2025-10-04T03:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.991395 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.991475 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.991492 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.991520 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:56 crc kubenswrapper[4742]: I1004 03:11:56.991537 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:56Z","lastTransitionTime":"2025-10-04T03:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.094548 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.094645 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.094666 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.094700 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.094720 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:57Z","lastTransitionTime":"2025-10-04T03:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.198603 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.198699 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.198723 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.198781 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.198801 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:57Z","lastTransitionTime":"2025-10-04T03:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.303121 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.303206 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.303224 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.303245 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.303259 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:57Z","lastTransitionTime":"2025-10-04T03:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.405876 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.406000 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.406013 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.406033 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.406046 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:57Z","lastTransitionTime":"2025-10-04T03:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.508840 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.508909 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.508927 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.508952 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.508970 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:57Z","lastTransitionTime":"2025-10-04T03:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.612523 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.612570 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.612580 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.612596 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.612606 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:57Z","lastTransitionTime":"2025-10-04T03:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.715928 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.715986 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.715999 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.716031 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.716049 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:57Z","lastTransitionTime":"2025-10-04T03:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.819497 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.819561 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.819578 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.819602 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.819619 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:57Z","lastTransitionTime":"2025-10-04T03:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.857746 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:11:57 crc kubenswrapper[4742]: E1004 03:11:57.857940 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.922756 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.922828 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.922840 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.922860 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:57 crc kubenswrapper[4742]: I1004 03:11:57.922873 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:57Z","lastTransitionTime":"2025-10-04T03:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.026431 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.026480 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.026488 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.026505 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.026531 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:58Z","lastTransitionTime":"2025-10-04T03:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.130460 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.130551 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.130575 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.130610 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.130634 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:58Z","lastTransitionTime":"2025-10-04T03:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.233733 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.233803 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.233822 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.233875 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.233888 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:58Z","lastTransitionTime":"2025-10-04T03:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.337405 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.337474 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.337490 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.337512 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.337528 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:58Z","lastTransitionTime":"2025-10-04T03:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.441724 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.441782 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.441814 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.441838 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.441922 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:58Z","lastTransitionTime":"2025-10-04T03:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.545709 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.545773 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.545792 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.545820 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.545839 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:58Z","lastTransitionTime":"2025-10-04T03:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.642004 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.642234 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:58 crc kubenswrapper[4742]: E1004 03:11:58.642392 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:13:02.642359666 +0000 UTC m=+148.574696882 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:11:58 crc kubenswrapper[4742]: E1004 03:11:58.642499 4742 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:11:58 crc kubenswrapper[4742]: E1004 03:11:58.642672 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:13:02.642636694 +0000 UTC m=+148.574973900 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.649895 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.649960 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.649979 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.650007 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.650024 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:58Z","lastTransitionTime":"2025-10-04T03:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.744080 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.744176 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.744245 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:11:58 crc kubenswrapper[4742]: E1004 03:11:58.744480 4742 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:11:58 crc kubenswrapper[4742]: E1004 03:11:58.744522 4742 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:11:58 crc kubenswrapper[4742]: E1004 03:11:58.744701 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:13:02.744663492 +0000 UTC m=+148.677000698 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:11:58 crc kubenswrapper[4742]: E1004 03:11:58.744518 4742 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:11:58 crc kubenswrapper[4742]: E1004 03:11:58.744774 4742 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:11:58 crc kubenswrapper[4742]: E1004 03:11:58.744799 4742 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:11:58 crc kubenswrapper[4742]: E1004 03:11:58.744560 4742 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:11:58 crc kubenswrapper[4742]: E1004 03:11:58.744895 4742 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:11:58 crc kubenswrapper[4742]: E1004 03:11:58.744898 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 03:13:02.744857067 +0000 UTC m=+148.677194283 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:11:58 crc kubenswrapper[4742]: E1004 03:11:58.745028 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 03:13:02.744999861 +0000 UTC m=+148.677337067 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.753804 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.753870 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.753893 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.753922 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.753943 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:58Z","lastTransitionTime":"2025-10-04T03:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.857379 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.857430 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.857384 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:11:58 crc kubenswrapper[4742]: E1004 03:11:58.857531 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.857539 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.857588 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.857601 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.857617 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.857631 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:58Z","lastTransitionTime":"2025-10-04T03:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:58 crc kubenswrapper[4742]: E1004 03:11:58.857655 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:11:58 crc kubenswrapper[4742]: E1004 03:11:58.857759 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.960671 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.960723 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.960733 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.960747 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:58 crc kubenswrapper[4742]: I1004 03:11:58.960756 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:58Z","lastTransitionTime":"2025-10-04T03:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.063957 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.063999 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.064008 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.064022 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.064032 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:59Z","lastTransitionTime":"2025-10-04T03:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.167174 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.167242 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.167261 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.167324 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.167344 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:59Z","lastTransitionTime":"2025-10-04T03:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.270478 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.270527 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.270569 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.270596 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.270610 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:59Z","lastTransitionTime":"2025-10-04T03:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.373755 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.373797 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.373809 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.373828 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.373842 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:59Z","lastTransitionTime":"2025-10-04T03:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.477675 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.477732 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.477752 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.477779 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.477801 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:59Z","lastTransitionTime":"2025-10-04T03:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.581452 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.581506 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.581516 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.581532 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.581544 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:59Z","lastTransitionTime":"2025-10-04T03:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.684666 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.684762 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.684779 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.684805 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.684820 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:59Z","lastTransitionTime":"2025-10-04T03:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.787642 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.787709 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.787732 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.787764 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.787787 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:59Z","lastTransitionTime":"2025-10-04T03:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.857645 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:11:59 crc kubenswrapper[4742]: E1004 03:11:59.857876 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.891203 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.891294 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.891311 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.891347 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.891361 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:59Z","lastTransitionTime":"2025-10-04T03:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.994485 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.994544 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.994562 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.994587 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:11:59 crc kubenswrapper[4742]: I1004 03:11:59.994605 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:11:59Z","lastTransitionTime":"2025-10-04T03:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.097243 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.097366 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.097386 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.097411 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.097430 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:00Z","lastTransitionTime":"2025-10-04T03:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.201400 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.201450 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.201470 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.201495 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.201517 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:00Z","lastTransitionTime":"2025-10-04T03:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.305044 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.305101 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.305111 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.305127 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.305136 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:00Z","lastTransitionTime":"2025-10-04T03:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.408677 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.408739 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.408759 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.408786 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.408806 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:00Z","lastTransitionTime":"2025-10-04T03:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.512362 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.512423 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.512438 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.512458 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.512471 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:00Z","lastTransitionTime":"2025-10-04T03:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.615627 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.615686 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.615703 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.615729 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.615749 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:00Z","lastTransitionTime":"2025-10-04T03:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.719741 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.719796 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.719807 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.719828 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.719841 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:00Z","lastTransitionTime":"2025-10-04T03:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.823480 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.823548 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.823566 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.823624 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.823651 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:00Z","lastTransitionTime":"2025-10-04T03:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.857627 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.857764 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.857867 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:12:00 crc kubenswrapper[4742]: E1004 03:12:00.857867 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:12:00 crc kubenswrapper[4742]: E1004 03:12:00.857952 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:12:00 crc kubenswrapper[4742]: E1004 03:12:00.858029 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.926804 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.926893 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.926917 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.926950 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:00 crc kubenswrapper[4742]: I1004 03:12:00.926971 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:00Z","lastTransitionTime":"2025-10-04T03:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.030098 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.030140 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.030154 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.030174 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.030187 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:01Z","lastTransitionTime":"2025-10-04T03:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.133112 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.133170 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.133189 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.133214 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.133234 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:01Z","lastTransitionTime":"2025-10-04T03:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.236262 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.236364 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.236380 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.236404 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.236424 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:01Z","lastTransitionTime":"2025-10-04T03:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.339502 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.339586 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.339599 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.339619 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.339630 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:01Z","lastTransitionTime":"2025-10-04T03:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.443074 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.443647 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.443666 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.443696 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.443716 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:01Z","lastTransitionTime":"2025-10-04T03:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.545877 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.545935 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.545950 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.545975 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.545991 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:01Z","lastTransitionTime":"2025-10-04T03:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.648436 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.648482 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.648496 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.648517 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.648532 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:01Z","lastTransitionTime":"2025-10-04T03:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.751752 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.751808 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.751818 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.751837 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.751851 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:01Z","lastTransitionTime":"2025-10-04T03:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.855047 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.855169 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.855184 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.855236 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.855252 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:01Z","lastTransitionTime":"2025-10-04T03:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.857387 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:12:01 crc kubenswrapper[4742]: E1004 03:12:01.857586 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.950130 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.950181 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.950191 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.950214 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.950225 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:01Z","lastTransitionTime":"2025-10-04T03:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:01 crc kubenswrapper[4742]: E1004 03:12:01.963511 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:12:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.968535 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.968591 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.968601 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.968618 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.968632 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:01Z","lastTransitionTime":"2025-10-04T03:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:01 crc kubenswrapper[4742]: E1004 03:12:01.982845 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:12:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.987983 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.988016 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.988028 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.988044 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:01 crc kubenswrapper[4742]: I1004 03:12:01.988057 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:01Z","lastTransitionTime":"2025-10-04T03:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:02 crc kubenswrapper[4742]: E1004 03:12:02.007634 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:12:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.013937 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.014018 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.014042 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.014071 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.014092 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:02Z","lastTransitionTime":"2025-10-04T03:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:02 crc kubenswrapper[4742]: E1004 03:12:02.034359 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:12:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.040060 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.040125 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.040170 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.040196 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.040208 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:02Z","lastTransitionTime":"2025-10-04T03:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:02 crc kubenswrapper[4742]: E1004 03:12:02.055308 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:12:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:12:02 crc kubenswrapper[4742]: E1004 03:12:02.055553 4742 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.057957 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.058020 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.058031 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.058050 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.058062 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:02Z","lastTransitionTime":"2025-10-04T03:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.161109 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.161162 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.161173 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.161195 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.161209 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:02Z","lastTransitionTime":"2025-10-04T03:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.264118 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.264173 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.264188 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.264210 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.264226 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:02Z","lastTransitionTime":"2025-10-04T03:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.367643 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.367710 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.367731 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.367779 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.367800 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:02Z","lastTransitionTime":"2025-10-04T03:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.471743 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.471795 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.471808 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.471831 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.471849 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:02Z","lastTransitionTime":"2025-10-04T03:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.576015 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.576115 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.576139 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.576213 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.576241 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:02Z","lastTransitionTime":"2025-10-04T03:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.679784 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.679845 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.679855 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.679875 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.679887 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:02Z","lastTransitionTime":"2025-10-04T03:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.782242 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.782317 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.782331 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.782350 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.782362 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:02Z","lastTransitionTime":"2025-10-04T03:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.857711 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.857737 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.857806 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:12:02 crc kubenswrapper[4742]: E1004 03:12:02.857873 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:12:02 crc kubenswrapper[4742]: E1004 03:12:02.857993 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:12:02 crc kubenswrapper[4742]: E1004 03:12:02.858127 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.885472 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.885518 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.885527 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.885543 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.885552 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:02Z","lastTransitionTime":"2025-10-04T03:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.989057 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.989142 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.989163 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.989189 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:02 crc kubenswrapper[4742]: I1004 03:12:02.989208 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:02Z","lastTransitionTime":"2025-10-04T03:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.091751 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.091818 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.091856 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.091888 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.091907 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:03Z","lastTransitionTime":"2025-10-04T03:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.199870 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.199921 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.199933 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.199953 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.199966 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:03Z","lastTransitionTime":"2025-10-04T03:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.303209 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.303288 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.303298 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.303314 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.303324 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:03Z","lastTransitionTime":"2025-10-04T03:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.406800 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.406861 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.406875 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.406899 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.406920 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:03Z","lastTransitionTime":"2025-10-04T03:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.510337 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.510430 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.510445 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.510470 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.510487 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:03Z","lastTransitionTime":"2025-10-04T03:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.613591 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.613647 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.613656 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.613674 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.613686 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:03Z","lastTransitionTime":"2025-10-04T03:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.717253 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.717340 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.717354 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.717373 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.717384 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:03Z","lastTransitionTime":"2025-10-04T03:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.820698 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.820789 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.820801 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.820823 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.820835 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:03Z","lastTransitionTime":"2025-10-04T03:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.857860 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:12:03 crc kubenswrapper[4742]: E1004 03:12:03.858075 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.923335 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.923393 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.923410 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.923432 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:03 crc kubenswrapper[4742]: I1004 03:12:03.923448 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:03Z","lastTransitionTime":"2025-10-04T03:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.026394 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.026513 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.026530 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.026552 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.026567 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:04Z","lastTransitionTime":"2025-10-04T03:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.129808 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.129879 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.129894 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.129913 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.129930 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:04Z","lastTransitionTime":"2025-10-04T03:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.233773 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.233875 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.233889 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.233908 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.233920 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:04Z","lastTransitionTime":"2025-10-04T03:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.337301 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.337362 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.337376 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.337396 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.337409 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:04Z","lastTransitionTime":"2025-10-04T03:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.440744 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.440800 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.440815 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.440834 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.440845 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:04Z","lastTransitionTime":"2025-10-04T03:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.544427 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.544509 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.544531 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.544560 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.544598 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:04Z","lastTransitionTime":"2025-10-04T03:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.648685 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.648743 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.648755 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.648774 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.648786 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:04Z","lastTransitionTime":"2025-10-04T03:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.752005 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.752058 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.752071 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.752090 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.752103 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:04Z","lastTransitionTime":"2025-10-04T03:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.854546 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.854623 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.854637 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.854655 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.854667 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:04Z","lastTransitionTime":"2025-10-04T03:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.857962 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.857964 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:12:04 crc kubenswrapper[4742]: E1004 03:12:04.858096 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.858252 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:12:04 crc kubenswrapper[4742]: E1004 03:12:04.858355 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:12:04 crc kubenswrapper[4742]: E1004 03:12:04.858419 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.880516 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e264f7035e2361d73ee37805509fdf11fdadf89d4f44d5bf743a77efc9e25cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://057f5d92b218d0dbe502abb94c64e48331da32bca242dd2fedab0ce801ad4c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:12:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.898435 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b51a285702a7ca9176efe50434d406d91eff40c4dac211e71cc4152cfae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:12:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.919116 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b37c3bfb97a4dc710d7664dfa0bc1cd0093419f21a8ca483ac1f38e695d7271\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b37c3bfb97a4dc710d7664dfa0bc1cd0093419f21a8ca483ac1f38e695d7271\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:11:49Z\\\",\\\"message\\\":\\\"_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1004 03:11:49.775768 6758 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-nc8m7\\\\nI1004 03:11:49.775649 6758 services_controller.go:434] Service openshift-machine-api/control-plane-machine-set-operator retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{control-plane-machine-set-operator openshift-machine-api ffd0ef27-d28d-43cc-90c8-0e8843e4c04c 4409 0 2025-02-23 05:12:21 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:control-plane-machine-set-operator] map[capability.openshift.io/name:MachineAPI exclude.release.openshift.io/internal-openshift-hosted:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:control-plane-machine-set-operator-tls service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0077282e7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:9443,TargetPort:{1 0 https},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{k8s-app: control-plane-machine-set-operator,},ClusterIP:10.21\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9qrtl_openshift-ovn-kubernetes(0cfc6e64-ba08-4c44-a18e-3b93ab792a6a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-txjzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9qrtl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:12:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.935231 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f74200f-5729-4e70-8b07-b885654c56e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7871a801b96fa8ef62e65a229e503e2914a008f35163b85cbffc99ffc543c37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35257b186eeb8544b8912f2a2c995eccd0549a4be771fa68a17b6489fc5dc7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35257b186eeb8544b8912f2a2c995eccd0549a4be771fa68a17b6489fc5dc7ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:12:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.951658 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:12:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.957775 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.957842 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.957856 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.957893 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.957906 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:04Z","lastTransitionTime":"2025-10-04T03:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.973593 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ll99m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c84d43-0375-42ca-b3b3-9cadda96fcbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd1d6092975dd4703ff4930e2eafaf464f221510fddeda0f977a73c548c305eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://724111471d5875362f6163217cb9d41c4b4b9261adc4af56cb3d776e93b4906f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cac4b91029667185f8f4042aa5697ec5a413d9e6ae49a40fde8a60be5a0490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c782bd2ffcfb89fc5068b985f275157b932c5d258001ab766bf6329d5d19905\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5c53bf7f93c9f95538f756cb8a6d490315ad38e5eeb928a93fd3f7c19a60ed7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://919452d4f0492ace420e5234af0feffbb2302839599caa9b0b04129fe2de90ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94159458d9516393c6fecd4c05210bc26a9da9a527907760f1cdb9a6bf15187a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:11:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:11:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9dgsv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ll99m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:12:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:12:04 crc kubenswrapper[4742]: I1004 03:12:04.988488 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-cwm6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09a5aee9-b917-473a-9196-5cde6553c72a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6cefd4be7726c02ec9a7b2db28ce61c5dd11b708a131e05e96595b51719209b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrdmr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-cwm6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:12:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.004979 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cd46x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0302950c-deb3-420a-b798-c99459e62b3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fls9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fls9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:12Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cd46x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:12:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.030051 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91bc9f85-8f09-4d65-a81f-c8e50b0fbdb1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://830fe4584483d987ccd84235d0d565579604c385ffc648b6a61868a7369a19c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65ba85e90248976f60214ebcda8e19348e9c373aa509ca521cb8e313acec84c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01061b91a5279d579507bc180cc59378d3729dc8d327f90a37f3e89331d6a38d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f53f64a2c0a1b813ae42baac78f5e639dfc576212d66b48480b0d3563120e12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://467455dfc35412e2c98ab78be938378877791ef8e6acd843795dd5e600495934\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48906a72bb35674544fa9d074de2f582da4cbad41e0472f6bea749cd6a7d860c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48906a72bb35674544fa9d074de2f582da4cbad41e0472f6bea749cd6a7d860c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9bcffac1019b9253a814cd8b0951819a40309057f15c54bf6ff4ee0fc491e61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c9bcffac1019b9253a814cd8b0951819a40309057f15c54bf6ff4ee0fc491e61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0fd66b11b8abc5ab89eb8355cba078dc46e2be3ee0f59afc37cdcee8e80115b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd66b11b8abc5ab89eb8355cba078dc46e2be3ee0f59afc37cdcee8e80115b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:12:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.044693 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"989ea247-7ac9-4ade-8dcb-2632fef4f83e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://261e8ace3d214aec7b737ea538ac0996cc11a5dae65eacc33dd3a3ea65ed3558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8s6xh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qdt7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:12:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.061261 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.061347 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.061362 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.061387 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.061403 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:05Z","lastTransitionTime":"2025-10-04T03:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.063013 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-shjbc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8e4f883-06de-41e5-83c1-90f55f41061c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e11b178c3523269e8b4c09a1ad6f86eaee1d8f19f5056840dfa24466248889a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8xlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-shjbc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:12:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.077467 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ec6cadd-fdb4-43c5-a898-f1bde3e9898d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf178a15abcb94eda5033b74ef132d9e204b33f4112152edf172e61999fc2c9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac618416ef312d2c051d88c0348fea230d73d3c8d6f4b1830b4b584ceae645c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2ede8f31552e98a3087393a27a4bab2c56f116829fef768b9c1dda8d6630142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3ee868ef16effadb77812f1673237d65d10493bd29dfcef5c7fe53da640d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce3ee868ef16effadb77812f1673237d65d10493bd29dfcef5c7fe53da640d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:12:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.096501 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b8d717a7-fae1-4a36-8d23-3a304edc9e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://440031d12658d13fc18aafd83017658ca530f9bccb76d1fbd3aa75a17a95bf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0e328329ad7896d5b76c652bd5b083c75f10f1e1be2eb6256bd6a3b32eb4d6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c40f92730ebb4aca75d2f7f3e22f7a955cd1bed142f3356a6705a39c64ef1fef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efb33723046c2083793ffc2f3b9602b35ba20695d2ddd6fb28b98405657c2d8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8091365b05223f764d0d7ae4db7c0907be2b722fde37c7d6a8122f4530d6e610\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1004 03:10:48.478717 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:10:48.479681 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1409933045/tls.crt::/tmp/serving-cert-1409933045/tls.key\\\\\\\"\\\\nI1004 03:10:54.322502 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:10:54.325597 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:10:54.325626 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:10:54.325650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:10:54.325658 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:10:54.336578 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:10:54.336629 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:10:54.336637 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1004 03:10:54.336622 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 03:10:54.336643 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:10:54.336697 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:10:54.336703 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:10:54.336707 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 03:10:54.338689 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec7c16da1e53eaf1d71990d9bd867125dc8f1efabfb210346867c40f84d06890\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b46acf0f45e7676d08011106b6d7625e94e1b7e4d7e0bc1867422e5915e86fa1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:12:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.112198 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:12:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.130239 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:12:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.148372 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nc8m7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46de52ee-226c-4713-bd7b-961f8f43f635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe5cd4a2153e5ae33beedc1cfdee7c39f57d2dc41a3d5e7d442df75d3c069db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:11:45Z\\\",\\\"message\\\":\\\"2025-10-04T03:11:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a40fc70a-b392-4173-9861-b7ae2db0575b\\\\n2025-10-04T03:11:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a40fc70a-b392-4173-9861-b7ae2db0575b to /host/opt/cni/bin/\\\\n2025-10-04T03:11:00Z [verbose] multus-daemon started\\\\n2025-10-04T03:11:00Z [verbose] Readiness Indicator file check\\\\n2025-10-04T03:11:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:10:59Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8pj2t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nc8m7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:12:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.163042 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2694d67-7fb5-446f-a590-6113e9131921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:11:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b29154af28bbb704eeb5fde7a5c44db624f1cb35d79942c62d5ca9a03e697b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1070bde4d35430f4c712d63ede34a428ec20a277ce0409dd3e69dead27c3f98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgssc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:11:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwzp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:12:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.164470 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.164533 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.164547 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.164565 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.164578 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:05Z","lastTransitionTime":"2025-10-04T03:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.181876 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a644867d-761b-47a7-903d-ec499532c079\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78607013cde35d08c93caba210001f63756f8adca1e2aa7184b674615cfd8ec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3ca0db4eead058b8845fc8fcd4a57eae646f7f8701ad6970b8ed4041627ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3283aada514e6cc7d8ab358f6fcb6fa2d36050d614b1cdbb917dccec03ceb51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4b4f3ec2d6fdbf18f16dbc0cd1b7c74207b5ca763c54dcac3ed1c60ed4196\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:10:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:12:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.204168 4742 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:10:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://769de9b8cefb734d8d294d0dc956a3b96a4e8db582ebe9ddfde3637a16ea65cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:12:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.268555 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.268640 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.268659 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.268691 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.268711 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:05Z","lastTransitionTime":"2025-10-04T03:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.372097 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.372147 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.372156 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.372173 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.372183 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:05Z","lastTransitionTime":"2025-10-04T03:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.475698 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.475746 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.475757 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.475780 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.475791 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:05Z","lastTransitionTime":"2025-10-04T03:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.579626 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.579702 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.579714 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.579736 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.579746 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:05Z","lastTransitionTime":"2025-10-04T03:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.682606 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.682922 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.683002 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.683122 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.683216 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:05Z","lastTransitionTime":"2025-10-04T03:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.787070 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.787145 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.787166 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.787197 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.787218 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:05Z","lastTransitionTime":"2025-10-04T03:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.858691 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:12:05 crc kubenswrapper[4742]: E1004 03:12:05.859261 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.859552 4742 scope.go:117] "RemoveContainer" containerID="9b37c3bfb97a4dc710d7664dfa0bc1cd0093419f21a8ca483ac1f38e695d7271" Oct 04 03:12:05 crc kubenswrapper[4742]: E1004 03:12:05.859766 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9qrtl_openshift-ovn-kubernetes(0cfc6e64-ba08-4c44-a18e-3b93ab792a6a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.889366 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.889416 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.889426 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.889447 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.889461 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:05Z","lastTransitionTime":"2025-10-04T03:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.992139 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.992189 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.992200 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.992219 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:05 crc kubenswrapper[4742]: I1004 03:12:05.992237 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:05Z","lastTransitionTime":"2025-10-04T03:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.095797 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.095853 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.095865 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.095900 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.095927 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:06Z","lastTransitionTime":"2025-10-04T03:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.199826 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.199895 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.199906 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.199928 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.199943 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:06Z","lastTransitionTime":"2025-10-04T03:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.303811 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.303884 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.303897 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.303921 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.303936 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:06Z","lastTransitionTime":"2025-10-04T03:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.406368 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.406426 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.406434 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.406452 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.406464 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:06Z","lastTransitionTime":"2025-10-04T03:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.509876 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.509979 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.510006 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.510039 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.510065 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:06Z","lastTransitionTime":"2025-10-04T03:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.612888 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.612951 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.612968 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.612990 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.613008 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:06Z","lastTransitionTime":"2025-10-04T03:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.716184 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.716234 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.716244 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.716261 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.716288 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:06Z","lastTransitionTime":"2025-10-04T03:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.818804 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.818858 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.818871 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.818891 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.818903 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:06Z","lastTransitionTime":"2025-10-04T03:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.858027 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.858081 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.858225 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:12:06 crc kubenswrapper[4742]: E1004 03:12:06.858216 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:12:06 crc kubenswrapper[4742]: E1004 03:12:06.858445 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:12:06 crc kubenswrapper[4742]: E1004 03:12:06.858493 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.922252 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.922345 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.922366 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.922390 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:06 crc kubenswrapper[4742]: I1004 03:12:06.922407 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:06Z","lastTransitionTime":"2025-10-04T03:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.025594 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.025684 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.025709 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.025745 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.025784 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:07Z","lastTransitionTime":"2025-10-04T03:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.128465 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.128531 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.128545 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.128565 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.128576 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:07Z","lastTransitionTime":"2025-10-04T03:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.232762 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.232860 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.232887 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.232976 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.233044 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:07Z","lastTransitionTime":"2025-10-04T03:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.339052 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.339145 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.339170 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.339206 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.339230 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:07Z","lastTransitionTime":"2025-10-04T03:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.442503 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.442598 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.442617 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.442644 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.442665 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:07Z","lastTransitionTime":"2025-10-04T03:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.546059 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.546118 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.546130 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.546149 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.546162 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:07Z","lastTransitionTime":"2025-10-04T03:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.648569 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.648609 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.648619 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.648635 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.648647 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:07Z","lastTransitionTime":"2025-10-04T03:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.751252 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.751364 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.751380 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.751398 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.751413 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:07Z","lastTransitionTime":"2025-10-04T03:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.854585 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.854659 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.854678 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.854705 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.854723 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:07Z","lastTransitionTime":"2025-10-04T03:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.857653 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:12:07 crc kubenswrapper[4742]: E1004 03:12:07.857815 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.958152 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.958214 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.958231 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.958261 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:07 crc kubenswrapper[4742]: I1004 03:12:07.958351 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:07Z","lastTransitionTime":"2025-10-04T03:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.060645 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.060699 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.060711 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.060731 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.060744 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:08Z","lastTransitionTime":"2025-10-04T03:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.164424 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.164474 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.164489 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.164509 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.164523 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:08Z","lastTransitionTime":"2025-10-04T03:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.268267 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.268347 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.268355 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.268370 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.268380 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:08Z","lastTransitionTime":"2025-10-04T03:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.371125 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.371178 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.371189 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.371206 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.371216 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:08Z","lastTransitionTime":"2025-10-04T03:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.474130 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.474200 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.474218 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.474240 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.474254 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:08Z","lastTransitionTime":"2025-10-04T03:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.577654 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.577730 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.577752 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.577780 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.577800 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:08Z","lastTransitionTime":"2025-10-04T03:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.680719 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.680790 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.680803 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.680823 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.680836 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:08Z","lastTransitionTime":"2025-10-04T03:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.784363 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.784415 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.784428 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.784445 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.784457 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:08Z","lastTransitionTime":"2025-10-04T03:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.858514 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.858715 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.858761 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:12:08 crc kubenswrapper[4742]: E1004 03:12:08.858910 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:12:08 crc kubenswrapper[4742]: E1004 03:12:08.859189 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:12:08 crc kubenswrapper[4742]: E1004 03:12:08.859437 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.887014 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.887049 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.887058 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.887073 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.887086 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:08Z","lastTransitionTime":"2025-10-04T03:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.990351 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.990440 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.990482 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.990520 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:08 crc kubenswrapper[4742]: I1004 03:12:08.990547 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:08Z","lastTransitionTime":"2025-10-04T03:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.093052 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.093107 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.093117 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.093135 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.093146 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:09Z","lastTransitionTime":"2025-10-04T03:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.200081 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.200131 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.200153 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.200181 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.200197 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:09Z","lastTransitionTime":"2025-10-04T03:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.302676 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.302749 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.302767 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.302796 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.302812 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:09Z","lastTransitionTime":"2025-10-04T03:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.405972 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.406017 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.406026 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.406043 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.406054 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:09Z","lastTransitionTime":"2025-10-04T03:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.509297 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.509385 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.509400 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.509421 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.509435 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:09Z","lastTransitionTime":"2025-10-04T03:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.611899 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.611965 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.611977 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.612000 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.612018 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:09Z","lastTransitionTime":"2025-10-04T03:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.715141 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.715204 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.715222 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.715247 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.715267 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:09Z","lastTransitionTime":"2025-10-04T03:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.819788 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.819863 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.819886 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.819925 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.819947 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:09Z","lastTransitionTime":"2025-10-04T03:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.857434 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:12:09 crc kubenswrapper[4742]: E1004 03:12:09.857612 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.922399 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.922451 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.922461 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.922478 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:09 crc kubenswrapper[4742]: I1004 03:12:09.922488 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:09Z","lastTransitionTime":"2025-10-04T03:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.025969 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.026050 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.026069 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.026097 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.026115 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:10Z","lastTransitionTime":"2025-10-04T03:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.128516 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.128565 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.128575 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.128594 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.128604 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:10Z","lastTransitionTime":"2025-10-04T03:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.231410 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.231472 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.231485 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.231505 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.231518 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:10Z","lastTransitionTime":"2025-10-04T03:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.334235 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.334315 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.334335 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.334356 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.334372 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:10Z","lastTransitionTime":"2025-10-04T03:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.437223 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.437317 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.437330 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.437356 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.437370 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:10Z","lastTransitionTime":"2025-10-04T03:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.540480 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.540548 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.540566 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.540594 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.540617 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:10Z","lastTransitionTime":"2025-10-04T03:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.644471 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.644589 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.644616 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.644650 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.644678 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:10Z","lastTransitionTime":"2025-10-04T03:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.746873 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.746936 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.746956 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.746977 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.747182 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:10Z","lastTransitionTime":"2025-10-04T03:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.850849 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.850901 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.850911 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.850930 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.850940 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:10Z","lastTransitionTime":"2025-10-04T03:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.858142 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:12:10 crc kubenswrapper[4742]: E1004 03:12:10.858329 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.858489 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:12:10 crc kubenswrapper[4742]: E1004 03:12:10.858651 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.858757 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:12:10 crc kubenswrapper[4742]: E1004 03:12:10.859253 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.956193 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.956570 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.956596 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.956674 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:10 crc kubenswrapper[4742]: I1004 03:12:10.957113 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:10Z","lastTransitionTime":"2025-10-04T03:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.061422 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.061502 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.061514 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.061532 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.061588 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:11Z","lastTransitionTime":"2025-10-04T03:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.164837 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.164887 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.164898 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.164915 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.164927 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:11Z","lastTransitionTime":"2025-10-04T03:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.268322 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.268368 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.268377 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.268395 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.268406 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:11Z","lastTransitionTime":"2025-10-04T03:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.372512 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.372564 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.372576 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.372597 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.372611 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:11Z","lastTransitionTime":"2025-10-04T03:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.475640 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.475688 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.475697 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.475712 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.475721 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:11Z","lastTransitionTime":"2025-10-04T03:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.578788 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.578856 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.578873 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.578902 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.578920 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:11Z","lastTransitionTime":"2025-10-04T03:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.681925 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.681971 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.681980 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.681997 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.682016 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:11Z","lastTransitionTime":"2025-10-04T03:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.785624 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.785718 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.785736 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.785757 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.785770 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:11Z","lastTransitionTime":"2025-10-04T03:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.858428 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:12:11 crc kubenswrapper[4742]: E1004 03:12:11.858603 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.889414 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.889503 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.889526 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.889560 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.889583 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:11Z","lastTransitionTime":"2025-10-04T03:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.992257 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.992328 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.992337 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.992354 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:11 crc kubenswrapper[4742]: I1004 03:12:11.992364 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:11Z","lastTransitionTime":"2025-10-04T03:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.065582 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.065637 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.065649 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.065668 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.065687 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:12Z","lastTransitionTime":"2025-10-04T03:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:12 crc kubenswrapper[4742]: E1004 03:12:12.079262 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:12:12Z is after 2025-08-24T17:21:41Z" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.084366 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.084433 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.084447 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.084468 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.084484 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:12Z","lastTransitionTime":"2025-10-04T03:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:12 crc kubenswrapper[4742]: E1004 03:12:12.098112 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:12:12Z is after 2025-08-24T17:21:41Z" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.103248 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.103324 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.103340 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.103359 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.103372 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:12Z","lastTransitionTime":"2025-10-04T03:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:12 crc kubenswrapper[4742]: E1004 03:12:12.125542 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:12:12Z is after 2025-08-24T17:21:41Z" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.130799 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.130854 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.130867 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.130889 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.130904 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:12Z","lastTransitionTime":"2025-10-04T03:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:12 crc kubenswrapper[4742]: E1004 03:12:12.147770 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:12:12Z is after 2025-08-24T17:21:41Z" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.153080 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.153355 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.153474 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.153594 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.153706 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:12Z","lastTransitionTime":"2025-10-04T03:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:12 crc kubenswrapper[4742]: E1004 03:12:12.168505 4742 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:12:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"94af7d13-529b-4a4d-a127-c697b50cfaf5\\\",\\\"systemUUID\\\":\\\"91a5b4ce-a0c3-485b-af77-0ed26accb46e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:12:12Z is after 2025-08-24T17:21:41Z" Oct 04 03:12:12 crc kubenswrapper[4742]: E1004 03:12:12.169144 4742 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.171210 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.171258 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.171285 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.171307 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.171324 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:12Z","lastTransitionTime":"2025-10-04T03:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.274074 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.274413 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.274587 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.274791 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.275180 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:12Z","lastTransitionTime":"2025-10-04T03:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.377535 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.377578 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.377589 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.377609 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.377620 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:12Z","lastTransitionTime":"2025-10-04T03:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.480373 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.480719 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.480855 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.480994 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.481136 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:12Z","lastTransitionTime":"2025-10-04T03:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.584383 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.584438 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.584450 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.584467 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.584478 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:12Z","lastTransitionTime":"2025-10-04T03:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.687133 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.687219 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.687237 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.687261 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.687321 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:12Z","lastTransitionTime":"2025-10-04T03:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.790630 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.790694 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.790713 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.790740 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.790759 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:12Z","lastTransitionTime":"2025-10-04T03:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.857952 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.858052 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:12:12 crc kubenswrapper[4742]: E1004 03:12:12.858402 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.858127 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:12:12 crc kubenswrapper[4742]: E1004 03:12:12.858644 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:12:12 crc kubenswrapper[4742]: E1004 03:12:12.858942 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.893547 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.893598 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.893611 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.893628 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.893641 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:12Z","lastTransitionTime":"2025-10-04T03:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.995967 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.996017 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.996029 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.996046 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:12 crc kubenswrapper[4742]: I1004 03:12:12.996059 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:12Z","lastTransitionTime":"2025-10-04T03:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.098303 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.098345 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.098358 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.098376 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.098388 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:13Z","lastTransitionTime":"2025-10-04T03:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.201946 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.202441 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.202556 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.202665 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.202764 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:13Z","lastTransitionTime":"2025-10-04T03:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.305767 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.306157 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.306430 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.306516 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.306585 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:13Z","lastTransitionTime":"2025-10-04T03:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.409054 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.409117 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.409132 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.409153 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.409167 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:13Z","lastTransitionTime":"2025-10-04T03:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.512507 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.512554 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.512565 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.512581 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.512591 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:13Z","lastTransitionTime":"2025-10-04T03:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.615957 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.616014 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.616029 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.616049 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.616061 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:13Z","lastTransitionTime":"2025-10-04T03:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.718430 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.718495 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.718506 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.718525 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.718536 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:13Z","lastTransitionTime":"2025-10-04T03:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.821635 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.821679 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.821690 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.821706 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.821720 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:13Z","lastTransitionTime":"2025-10-04T03:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.857356 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:12:13 crc kubenswrapper[4742]: E1004 03:12:13.857670 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.924511 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.924575 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.924593 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.924618 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:13 crc kubenswrapper[4742]: I1004 03:12:13.924638 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:13Z","lastTransitionTime":"2025-10-04T03:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.027507 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.027558 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.027571 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.027592 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.027606 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:14Z","lastTransitionTime":"2025-10-04T03:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.130622 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.130687 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.130705 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.130731 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.130746 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:14Z","lastTransitionTime":"2025-10-04T03:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.233950 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.234019 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.234039 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.234064 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.234083 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:14Z","lastTransitionTime":"2025-10-04T03:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.337623 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.337692 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.337716 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.337746 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.337772 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:14Z","lastTransitionTime":"2025-10-04T03:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.441136 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.441204 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.441222 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.441252 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.441308 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:14Z","lastTransitionTime":"2025-10-04T03:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.544794 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.544956 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.544969 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.545021 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.545036 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:14Z","lastTransitionTime":"2025-10-04T03:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.648259 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.648345 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.648364 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.648388 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.648409 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:14Z","lastTransitionTime":"2025-10-04T03:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.751916 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.751970 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.751987 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.752011 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.752029 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:14Z","lastTransitionTime":"2025-10-04T03:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.855021 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.855055 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.855064 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.855083 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.855095 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:14Z","lastTransitionTime":"2025-10-04T03:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.857531 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.857621 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:12:14 crc kubenswrapper[4742]: E1004 03:12:14.857655 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:12:14 crc kubenswrapper[4742]: E1004 03:12:14.857885 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.857917 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:12:14 crc kubenswrapper[4742]: E1004 03:12:14.858033 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.920679 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=50.920639595 podStartE2EDuration="50.920639595s" podCreationTimestamp="2025-10-04 03:11:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:12:14.892586985 +0000 UTC m=+100.824924191" watchObservedRunningTime="2025-10-04 03:12:14.920639595 +0000 UTC m=+100.852976771" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.934700 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=80.934668611 podStartE2EDuration="1m20.934668611s" podCreationTimestamp="2025-10-04 03:10:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:12:14.921789715 +0000 UTC m=+100.854126891" watchObservedRunningTime="2025-10-04 03:12:14.934668611 +0000 UTC m=+100.867005817" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.953674 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-shjbc" podStartSLOduration=77.953649076 podStartE2EDuration="1m17.953649076s" podCreationTimestamp="2025-10-04 03:10:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:12:14.935496213 +0000 UTC m=+100.867833429" watchObservedRunningTime="2025-10-04 03:12:14.953649076 +0000 UTC m=+100.885986262" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.958645 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.958692 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.958705 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.958724 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.958740 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:14Z","lastTransitionTime":"2025-10-04T03:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:14 crc kubenswrapper[4742]: I1004 03:12:14.975534 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-nc8m7" podStartSLOduration=77.975504775 podStartE2EDuration="1m17.975504775s" podCreationTimestamp="2025-10-04 03:10:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:12:14.974851598 +0000 UTC m=+100.907188774" watchObservedRunningTime="2025-10-04 03:12:14.975504775 +0000 UTC m=+100.907841951" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.015566 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwzp4" podStartSLOduration=77.015542938 podStartE2EDuration="1m17.015542938s" podCreationTimestamp="2025-10-04 03:10:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:12:14.990749162 +0000 UTC m=+100.923086338" watchObservedRunningTime="2025-10-04 03:12:15.015542938 +0000 UTC m=+100.947880114" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.015778 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=80.015775695 podStartE2EDuration="1m20.015775695s" podCreationTimestamp="2025-10-04 03:10:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:12:15.014811259 +0000 UTC m=+100.947148455" watchObservedRunningTime="2025-10-04 03:12:15.015775695 +0000 UTC m=+100.948112871" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.061336 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.061370 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.061378 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.061394 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.061404 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:15Z","lastTransitionTime":"2025-10-04T03:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.113017 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=21.112992248 podStartE2EDuration="21.112992248s" podCreationTimestamp="2025-10-04 03:11:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:12:15.112565707 +0000 UTC m=+101.044902903" watchObservedRunningTime="2025-10-04 03:12:15.112992248 +0000 UTC m=+101.045329424" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.164094 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.164154 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.164169 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.164188 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.164201 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:15Z","lastTransitionTime":"2025-10-04T03:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.175444 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-cwm6r" podStartSLOduration=78.175412434 podStartE2EDuration="1m18.175412434s" podCreationTimestamp="2025-10-04 03:10:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:12:15.16104275 +0000 UTC m=+101.093379936" watchObservedRunningTime="2025-10-04 03:12:15.175412434 +0000 UTC m=+101.107749620" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.206343 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=22.206309669 podStartE2EDuration="22.206309669s" podCreationTimestamp="2025-10-04 03:11:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:12:15.205541989 +0000 UTC m=+101.137879225" watchObservedRunningTime="2025-10-04 03:12:15.206309669 +0000 UTC m=+101.138646865" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.253881 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podStartSLOduration=78.253844698 podStartE2EDuration="1m18.253844698s" podCreationTimestamp="2025-10-04 03:10:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:12:15.230993813 +0000 UTC m=+101.163331019" watchObservedRunningTime="2025-10-04 03:12:15.253844698 +0000 UTC m=+101.186181874" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.254863 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-ll99m" podStartSLOduration=78.254858564 podStartE2EDuration="1m18.254858564s" podCreationTimestamp="2025-10-04 03:10:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:12:15.253662043 +0000 UTC m=+101.185999239" watchObservedRunningTime="2025-10-04 03:12:15.254858564 +0000 UTC m=+101.187195740" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.267342 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.267406 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.267421 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.267443 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.267457 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:15Z","lastTransitionTime":"2025-10-04T03:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.370432 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.370489 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.370501 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.370521 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.370535 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:15Z","lastTransitionTime":"2025-10-04T03:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.473960 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.474009 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.474025 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.474043 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.474057 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:15Z","lastTransitionTime":"2025-10-04T03:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.576425 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.576494 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.576509 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.576528 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.576540 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:15Z","lastTransitionTime":"2025-10-04T03:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.680361 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.680445 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.680463 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.680492 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.680510 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:15Z","lastTransitionTime":"2025-10-04T03:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.783706 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.783772 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.783800 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.783834 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.783857 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:15Z","lastTransitionTime":"2025-10-04T03:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.858205 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:12:15 crc kubenswrapper[4742]: E1004 03:12:15.858451 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.887380 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.887432 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.887448 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.887471 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.887488 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:15Z","lastTransitionTime":"2025-10-04T03:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.991177 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.991250 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.991310 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.991350 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:15 crc kubenswrapper[4742]: I1004 03:12:15.991379 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:15Z","lastTransitionTime":"2025-10-04T03:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.094138 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.094223 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.094239 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.094296 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.094312 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:16Z","lastTransitionTime":"2025-10-04T03:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.196931 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.196976 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.196985 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.197030 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.197042 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:16Z","lastTransitionTime":"2025-10-04T03:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.300187 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.300236 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.300249 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.300290 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.300305 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:16Z","lastTransitionTime":"2025-10-04T03:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.403128 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.403184 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.403197 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.403217 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.403229 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:16Z","lastTransitionTime":"2025-10-04T03:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.506077 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.506153 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.506170 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.506201 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.506221 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:16Z","lastTransitionTime":"2025-10-04T03:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.609886 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.609954 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.609972 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.610001 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.610018 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:16Z","lastTransitionTime":"2025-10-04T03:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.664319 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0302950c-deb3-420a-b798-c99459e62b3f-metrics-certs\") pod \"network-metrics-daemon-cd46x\" (UID: \"0302950c-deb3-420a-b798-c99459e62b3f\") " pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:12:16 crc kubenswrapper[4742]: E1004 03:12:16.664544 4742 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:12:16 crc kubenswrapper[4742]: E1004 03:12:16.664681 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0302950c-deb3-420a-b798-c99459e62b3f-metrics-certs podName:0302950c-deb3-420a-b798-c99459e62b3f nodeName:}" failed. No retries permitted until 2025-10-04 03:13:20.664645919 +0000 UTC m=+166.596983135 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0302950c-deb3-420a-b798-c99459e62b3f-metrics-certs") pod "network-metrics-daemon-cd46x" (UID: "0302950c-deb3-420a-b798-c99459e62b3f") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.713962 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.714027 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.714045 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.714069 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.714086 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:16Z","lastTransitionTime":"2025-10-04T03:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.818064 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.818143 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.818159 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.818187 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.818206 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:16Z","lastTransitionTime":"2025-10-04T03:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.857733 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.857827 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:12:16 crc kubenswrapper[4742]: E1004 03:12:16.857915 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.857994 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:12:16 crc kubenswrapper[4742]: E1004 03:12:16.858142 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:12:16 crc kubenswrapper[4742]: E1004 03:12:16.858334 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.922099 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.922143 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.922155 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.922174 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:16 crc kubenswrapper[4742]: I1004 03:12:16.922187 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:16Z","lastTransitionTime":"2025-10-04T03:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.025415 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.025527 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.025545 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.025572 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.025591 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:17Z","lastTransitionTime":"2025-10-04T03:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.128473 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.128539 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.128550 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.128568 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.128582 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:17Z","lastTransitionTime":"2025-10-04T03:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.233056 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.233121 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.233135 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.233157 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.233172 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:17Z","lastTransitionTime":"2025-10-04T03:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.336580 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.336642 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.336665 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.336692 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.336714 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:17Z","lastTransitionTime":"2025-10-04T03:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.439364 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.439440 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.439460 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.439487 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.439507 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:17Z","lastTransitionTime":"2025-10-04T03:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.542720 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.542878 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.542905 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.542938 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.542964 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:17Z","lastTransitionTime":"2025-10-04T03:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.646551 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.646614 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.646627 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.646650 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.646667 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:17Z","lastTransitionTime":"2025-10-04T03:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.750201 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.750336 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.750367 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.750398 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.750421 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:17Z","lastTransitionTime":"2025-10-04T03:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.852873 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.852955 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.852975 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.853001 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.853019 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:17Z","lastTransitionTime":"2025-10-04T03:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.858183 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:12:17 crc kubenswrapper[4742]: E1004 03:12:17.858327 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.956729 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.956792 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.956805 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.956828 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:17 crc kubenswrapper[4742]: I1004 03:12:17.956848 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:17Z","lastTransitionTime":"2025-10-04T03:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.061000 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.061083 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.061101 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.061136 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.061159 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:18Z","lastTransitionTime":"2025-10-04T03:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.164451 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.164511 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.164523 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.164546 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.164559 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:18Z","lastTransitionTime":"2025-10-04T03:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.268460 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.268542 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.268563 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.268592 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.268614 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:18Z","lastTransitionTime":"2025-10-04T03:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.371426 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.371517 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.371542 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.371586 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.371616 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:18Z","lastTransitionTime":"2025-10-04T03:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.475014 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.475085 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.475121 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.475159 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.475179 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:18Z","lastTransitionTime":"2025-10-04T03:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.578751 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.578825 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.578845 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.578871 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.578893 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:18Z","lastTransitionTime":"2025-10-04T03:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.683208 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.683292 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.683309 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.683330 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.683343 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:18Z","lastTransitionTime":"2025-10-04T03:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.786898 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.786960 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.786977 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.787001 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.787018 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:18Z","lastTransitionTime":"2025-10-04T03:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.858187 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.858244 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.858263 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:12:18 crc kubenswrapper[4742]: E1004 03:12:18.858500 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:12:18 crc kubenswrapper[4742]: E1004 03:12:18.858728 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:12:18 crc kubenswrapper[4742]: E1004 03:12:18.858904 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.889994 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.890052 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.890072 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.890095 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.890114 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:18Z","lastTransitionTime":"2025-10-04T03:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.993560 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.993630 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.993650 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.993678 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:18 crc kubenswrapper[4742]: I1004 03:12:18.993699 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:18Z","lastTransitionTime":"2025-10-04T03:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.097392 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.097491 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.097516 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.097556 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.097582 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:19Z","lastTransitionTime":"2025-10-04T03:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.200444 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.200541 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.200561 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.200593 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.200611 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:19Z","lastTransitionTime":"2025-10-04T03:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.303812 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.303865 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.303878 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.303895 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.303909 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:19Z","lastTransitionTime":"2025-10-04T03:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.410194 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.410252 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.410262 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.410312 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.410329 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:19Z","lastTransitionTime":"2025-10-04T03:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.513326 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.513584 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.513652 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.513760 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.513833 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:19Z","lastTransitionTime":"2025-10-04T03:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.616860 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.617169 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.617306 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.617388 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.617481 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:19Z","lastTransitionTime":"2025-10-04T03:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.720442 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.720538 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.720563 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.720597 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.720618 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:19Z","lastTransitionTime":"2025-10-04T03:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.824450 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.824532 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.824552 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.824585 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.824605 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:19Z","lastTransitionTime":"2025-10-04T03:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.857947 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:12:19 crc kubenswrapper[4742]: E1004 03:12:19.858177 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.928256 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.928333 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.928345 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.928364 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:19 crc kubenswrapper[4742]: I1004 03:12:19.928376 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:19Z","lastTransitionTime":"2025-10-04T03:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.030707 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.030748 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.030759 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.030780 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.030792 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:20Z","lastTransitionTime":"2025-10-04T03:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.133777 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.134108 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.134186 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.134292 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.134411 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:20Z","lastTransitionTime":"2025-10-04T03:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.237159 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.237248 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.237319 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.237358 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.237383 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:20Z","lastTransitionTime":"2025-10-04T03:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.340659 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.340708 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.340720 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.340741 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.340755 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:20Z","lastTransitionTime":"2025-10-04T03:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.444126 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.444179 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.444191 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.444210 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.444225 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:20Z","lastTransitionTime":"2025-10-04T03:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.547692 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.547764 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.547785 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.547814 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.547833 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:20Z","lastTransitionTime":"2025-10-04T03:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.651359 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.651417 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.651434 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.651455 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.651470 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:20Z","lastTransitionTime":"2025-10-04T03:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.754686 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.754746 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.754760 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.754784 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.754799 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:20Z","lastTransitionTime":"2025-10-04T03:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.857666 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.857728 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:12:20 crc kubenswrapper[4742]: E1004 03:12:20.857920 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.858024 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.858051 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.858073 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.858085 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.858100 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.858111 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:20Z","lastTransitionTime":"2025-10-04T03:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:20 crc kubenswrapper[4742]: E1004 03:12:20.858555 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:12:20 crc kubenswrapper[4742]: E1004 03:12:20.858691 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.858849 4742 scope.go:117] "RemoveContainer" containerID="9b37c3bfb97a4dc710d7664dfa0bc1cd0093419f21a8ca483ac1f38e695d7271" Oct 04 03:12:20 crc kubenswrapper[4742]: E1004 03:12:20.859037 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9qrtl_openshift-ovn-kubernetes(0cfc6e64-ba08-4c44-a18e-3b93ab792a6a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.960330 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.960400 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.960416 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.960435 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:20 crc kubenswrapper[4742]: I1004 03:12:20.960452 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:20Z","lastTransitionTime":"2025-10-04T03:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.063160 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.063237 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.063255 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.063327 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.063356 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:21Z","lastTransitionTime":"2025-10-04T03:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.165844 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.165901 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.165914 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.165934 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.165946 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:21Z","lastTransitionTime":"2025-10-04T03:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.268972 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.269038 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.269060 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.269093 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.269115 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:21Z","lastTransitionTime":"2025-10-04T03:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.371778 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.371823 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.371834 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.371872 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.371883 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:21Z","lastTransitionTime":"2025-10-04T03:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.474688 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.474946 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.475042 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.475136 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.475221 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:21Z","lastTransitionTime":"2025-10-04T03:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.578871 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.579186 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.579286 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.579355 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.579429 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:21Z","lastTransitionTime":"2025-10-04T03:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.682431 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.682475 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.682484 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.682502 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.682512 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:21Z","lastTransitionTime":"2025-10-04T03:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.785065 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.785121 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.785136 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.785157 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.785171 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:21Z","lastTransitionTime":"2025-10-04T03:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.857610 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:12:21 crc kubenswrapper[4742]: E1004 03:12:21.857802 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.887476 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.887519 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.887531 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.887550 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.887563 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:21Z","lastTransitionTime":"2025-10-04T03:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.990258 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.990342 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.990356 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.990376 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:21 crc kubenswrapper[4742]: I1004 03:12:21.990390 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:21Z","lastTransitionTime":"2025-10-04T03:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.093737 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.093793 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.093805 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.093824 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.093875 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:22Z","lastTransitionTime":"2025-10-04T03:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.195984 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.196028 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.196037 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.196051 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.196059 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:22Z","lastTransitionTime":"2025-10-04T03:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.301076 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.301141 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.301157 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.301191 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.301206 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:22Z","lastTransitionTime":"2025-10-04T03:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.404165 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.404228 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.404240 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.404290 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.404309 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:22Z","lastTransitionTime":"2025-10-04T03:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.507492 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.507531 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.507543 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.507561 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.507573 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:22Z","lastTransitionTime":"2025-10-04T03:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.550038 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.550091 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.550105 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.550123 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.550387 4742 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:12:22Z","lastTransitionTime":"2025-10-04T03:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.588893 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-vmqjk"] Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.589454 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vmqjk" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.591167 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.591842 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.592012 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.592080 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.631643 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/6927ab03-f3a9-4cbd-a13f-099bfd7d7517-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-vmqjk\" (UID: \"6927ab03-f3a9-4cbd-a13f-099bfd7d7517\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vmqjk" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.631689 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6927ab03-f3a9-4cbd-a13f-099bfd7d7517-service-ca\") pod \"cluster-version-operator-5c965bbfc6-vmqjk\" (UID: \"6927ab03-f3a9-4cbd-a13f-099bfd7d7517\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vmqjk" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.631720 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/6927ab03-f3a9-4cbd-a13f-099bfd7d7517-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-vmqjk\" (UID: \"6927ab03-f3a9-4cbd-a13f-099bfd7d7517\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vmqjk" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.631758 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6927ab03-f3a9-4cbd-a13f-099bfd7d7517-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-vmqjk\" (UID: \"6927ab03-f3a9-4cbd-a13f-099bfd7d7517\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vmqjk" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.631815 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6927ab03-f3a9-4cbd-a13f-099bfd7d7517-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-vmqjk\" (UID: \"6927ab03-f3a9-4cbd-a13f-099bfd7d7517\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vmqjk" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.732745 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6927ab03-f3a9-4cbd-a13f-099bfd7d7517-service-ca\") pod \"cluster-version-operator-5c965bbfc6-vmqjk\" (UID: \"6927ab03-f3a9-4cbd-a13f-099bfd7d7517\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vmqjk" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.732830 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/6927ab03-f3a9-4cbd-a13f-099bfd7d7517-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-vmqjk\" (UID: \"6927ab03-f3a9-4cbd-a13f-099bfd7d7517\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vmqjk" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.732855 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6927ab03-f3a9-4cbd-a13f-099bfd7d7517-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-vmqjk\" (UID: \"6927ab03-f3a9-4cbd-a13f-099bfd7d7517\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vmqjk" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.732884 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6927ab03-f3a9-4cbd-a13f-099bfd7d7517-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-vmqjk\" (UID: \"6927ab03-f3a9-4cbd-a13f-099bfd7d7517\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vmqjk" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.732926 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/6927ab03-f3a9-4cbd-a13f-099bfd7d7517-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-vmqjk\" (UID: \"6927ab03-f3a9-4cbd-a13f-099bfd7d7517\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vmqjk" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.732938 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/6927ab03-f3a9-4cbd-a13f-099bfd7d7517-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-vmqjk\" (UID: \"6927ab03-f3a9-4cbd-a13f-099bfd7d7517\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vmqjk" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.732991 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/6927ab03-f3a9-4cbd-a13f-099bfd7d7517-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-vmqjk\" (UID: \"6927ab03-f3a9-4cbd-a13f-099bfd7d7517\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vmqjk" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.734822 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6927ab03-f3a9-4cbd-a13f-099bfd7d7517-service-ca\") pod \"cluster-version-operator-5c965bbfc6-vmqjk\" (UID: \"6927ab03-f3a9-4cbd-a13f-099bfd7d7517\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vmqjk" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.738419 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6927ab03-f3a9-4cbd-a13f-099bfd7d7517-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-vmqjk\" (UID: \"6927ab03-f3a9-4cbd-a13f-099bfd7d7517\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vmqjk" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.749765 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6927ab03-f3a9-4cbd-a13f-099bfd7d7517-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-vmqjk\" (UID: \"6927ab03-f3a9-4cbd-a13f-099bfd7d7517\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vmqjk" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.858206 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:12:22 crc kubenswrapper[4742]: E1004 03:12:22.858407 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.858526 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:12:22 crc kubenswrapper[4742]: E1004 03:12:22.858697 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.858778 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:12:22 crc kubenswrapper[4742]: E1004 03:12:22.858834 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:12:22 crc kubenswrapper[4742]: I1004 03:12:22.908020 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vmqjk" Oct 04 03:12:23 crc kubenswrapper[4742]: I1004 03:12:23.426448 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vmqjk" event={"ID":"6927ab03-f3a9-4cbd-a13f-099bfd7d7517","Type":"ContainerStarted","Data":"4ab67d10ac9b518a75d5deaf2d66676a28979b27520e37bb732f610df19522be"} Oct 04 03:12:23 crc kubenswrapper[4742]: I1004 03:12:23.426894 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vmqjk" event={"ID":"6927ab03-f3a9-4cbd-a13f-099bfd7d7517","Type":"ContainerStarted","Data":"ce0ce41f3f9d8d3ca9c8f8510a327845ca2cad40c6d7c656db628b85645f1903"} Oct 04 03:12:23 crc kubenswrapper[4742]: I1004 03:12:23.857810 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:12:23 crc kubenswrapper[4742]: E1004 03:12:23.858005 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:12:24 crc kubenswrapper[4742]: I1004 03:12:24.858417 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:12:24 crc kubenswrapper[4742]: E1004 03:12:24.859348 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:12:24 crc kubenswrapper[4742]: I1004 03:12:24.859365 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:12:24 crc kubenswrapper[4742]: E1004 03:12:24.859479 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:12:24 crc kubenswrapper[4742]: I1004 03:12:24.859393 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:12:24 crc kubenswrapper[4742]: E1004 03:12:24.859535 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:12:25 crc kubenswrapper[4742]: I1004 03:12:25.857548 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:12:25 crc kubenswrapper[4742]: E1004 03:12:25.857814 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:12:26 crc kubenswrapper[4742]: I1004 03:12:26.858072 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:12:26 crc kubenswrapper[4742]: E1004 03:12:26.858388 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:12:26 crc kubenswrapper[4742]: I1004 03:12:26.858824 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:12:26 crc kubenswrapper[4742]: I1004 03:12:26.858906 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:12:26 crc kubenswrapper[4742]: E1004 03:12:26.858977 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:12:26 crc kubenswrapper[4742]: E1004 03:12:26.858969 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:12:27 crc kubenswrapper[4742]: I1004 03:12:27.857813 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:12:27 crc kubenswrapper[4742]: E1004 03:12:27.858053 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:12:28 crc kubenswrapper[4742]: I1004 03:12:28.857786 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:12:28 crc kubenswrapper[4742]: I1004 03:12:28.857890 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:12:28 crc kubenswrapper[4742]: E1004 03:12:28.857956 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:12:28 crc kubenswrapper[4742]: I1004 03:12:28.857786 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:12:28 crc kubenswrapper[4742]: E1004 03:12:28.858063 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:12:28 crc kubenswrapper[4742]: E1004 03:12:28.858112 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:12:29 crc kubenswrapper[4742]: I1004 03:12:29.858403 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:12:29 crc kubenswrapper[4742]: E1004 03:12:29.859234 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:12:30 crc kubenswrapper[4742]: I1004 03:12:30.857623 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:12:30 crc kubenswrapper[4742]: I1004 03:12:30.857648 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:12:30 crc kubenswrapper[4742]: I1004 03:12:30.857682 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:12:30 crc kubenswrapper[4742]: E1004 03:12:30.858239 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:12:30 crc kubenswrapper[4742]: E1004 03:12:30.858594 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:12:30 crc kubenswrapper[4742]: E1004 03:12:30.858772 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:12:31 crc kubenswrapper[4742]: I1004 03:12:31.857660 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:12:31 crc kubenswrapper[4742]: E1004 03:12:31.858413 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:12:32 crc kubenswrapper[4742]: I1004 03:12:32.463659 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nc8m7_46de52ee-226c-4713-bd7b-961f8f43f635/kube-multus/1.log" Oct 04 03:12:32 crc kubenswrapper[4742]: I1004 03:12:32.464462 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nc8m7_46de52ee-226c-4713-bd7b-961f8f43f635/kube-multus/0.log" Oct 04 03:12:32 crc kubenswrapper[4742]: I1004 03:12:32.464547 4742 generic.go:334] "Generic (PLEG): container finished" podID="46de52ee-226c-4713-bd7b-961f8f43f635" containerID="bfe5cd4a2153e5ae33beedc1cfdee7c39f57d2dc41a3d5e7d442df75d3c069db" exitCode=1 Oct 04 03:12:32 crc kubenswrapper[4742]: I1004 03:12:32.464602 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nc8m7" event={"ID":"46de52ee-226c-4713-bd7b-961f8f43f635","Type":"ContainerDied","Data":"bfe5cd4a2153e5ae33beedc1cfdee7c39f57d2dc41a3d5e7d442df75d3c069db"} Oct 04 03:12:32 crc kubenswrapper[4742]: I1004 03:12:32.464658 4742 scope.go:117] "RemoveContainer" containerID="7a52af85f106dc6223a0c6b63c6c286cda4fa538ca704351ab5cdbeca2088b8c" Oct 04 03:12:32 crc kubenswrapper[4742]: I1004 03:12:32.465740 4742 scope.go:117] "RemoveContainer" containerID="bfe5cd4a2153e5ae33beedc1cfdee7c39f57d2dc41a3d5e7d442df75d3c069db" Oct 04 03:12:32 crc kubenswrapper[4742]: E1004 03:12:32.466135 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-nc8m7_openshift-multus(46de52ee-226c-4713-bd7b-961f8f43f635)\"" pod="openshift-multus/multus-nc8m7" podUID="46de52ee-226c-4713-bd7b-961f8f43f635" Oct 04 03:12:32 crc kubenswrapper[4742]: I1004 03:12:32.487516 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vmqjk" podStartSLOduration=95.48748968 podStartE2EDuration="1m35.48748968s" podCreationTimestamp="2025-10-04 03:10:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:12:23.450194142 +0000 UTC m=+109.382531368" watchObservedRunningTime="2025-10-04 03:12:32.48748968 +0000 UTC m=+118.419826856" Oct 04 03:12:32 crc kubenswrapper[4742]: I1004 03:12:32.857924 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:12:32 crc kubenswrapper[4742]: I1004 03:12:32.857989 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:12:32 crc kubenswrapper[4742]: I1004 03:12:32.858072 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:12:32 crc kubenswrapper[4742]: E1004 03:12:32.858336 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:12:32 crc kubenswrapper[4742]: E1004 03:12:32.858491 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:12:32 crc kubenswrapper[4742]: E1004 03:12:32.858755 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:12:33 crc kubenswrapper[4742]: I1004 03:12:33.468774 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nc8m7_46de52ee-226c-4713-bd7b-961f8f43f635/kube-multus/1.log" Oct 04 03:12:33 crc kubenswrapper[4742]: I1004 03:12:33.857824 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:12:33 crc kubenswrapper[4742]: E1004 03:12:33.858024 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:12:33 crc kubenswrapper[4742]: I1004 03:12:33.858988 4742 scope.go:117] "RemoveContainer" containerID="9b37c3bfb97a4dc710d7664dfa0bc1cd0093419f21a8ca483ac1f38e695d7271" Oct 04 03:12:34 crc kubenswrapper[4742]: I1004 03:12:34.474384 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qrtl_0cfc6e64-ba08-4c44-a18e-3b93ab792a6a/ovnkube-controller/3.log" Oct 04 03:12:34 crc kubenswrapper[4742]: I1004 03:12:34.477114 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" event={"ID":"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a","Type":"ContainerStarted","Data":"8bf4cc53b1c45b02faf41e5d40191f4fc7ff1959030bfa049955d8e7a31350e7"} Oct 04 03:12:34 crc kubenswrapper[4742]: I1004 03:12:34.477706 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:12:34 crc kubenswrapper[4742]: I1004 03:12:34.513782 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" podStartSLOduration=97.513762978 podStartE2EDuration="1m37.513762978s" podCreationTimestamp="2025-10-04 03:10:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:12:34.513561353 +0000 UTC m=+120.445898549" watchObservedRunningTime="2025-10-04 03:12:34.513762978 +0000 UTC m=+120.446100164" Oct 04 03:12:34 crc kubenswrapper[4742]: I1004 03:12:34.821587 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-cd46x"] Oct 04 03:12:34 crc kubenswrapper[4742]: I1004 03:12:34.821739 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:12:34 crc kubenswrapper[4742]: E1004 03:12:34.821843 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:12:34 crc kubenswrapper[4742]: E1004 03:12:34.851894 4742 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 04 03:12:34 crc kubenswrapper[4742]: I1004 03:12:34.858755 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:12:34 crc kubenswrapper[4742]: I1004 03:12:34.858853 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:12:34 crc kubenswrapper[4742]: E1004 03:12:34.858983 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:12:34 crc kubenswrapper[4742]: I1004 03:12:34.859120 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:12:34 crc kubenswrapper[4742]: E1004 03:12:34.859200 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:12:34 crc kubenswrapper[4742]: E1004 03:12:34.859258 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:12:34 crc kubenswrapper[4742]: E1004 03:12:34.964937 4742 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 03:12:36 crc kubenswrapper[4742]: I1004 03:12:36.857406 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:12:36 crc kubenswrapper[4742]: I1004 03:12:36.857576 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:12:36 crc kubenswrapper[4742]: E1004 03:12:36.857820 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:12:36 crc kubenswrapper[4742]: I1004 03:12:36.857693 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:12:36 crc kubenswrapper[4742]: I1004 03:12:36.857652 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:12:36 crc kubenswrapper[4742]: E1004 03:12:36.857971 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:12:36 crc kubenswrapper[4742]: E1004 03:12:36.858362 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:12:36 crc kubenswrapper[4742]: E1004 03:12:36.858462 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:12:38 crc kubenswrapper[4742]: I1004 03:12:38.858297 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:12:38 crc kubenswrapper[4742]: I1004 03:12:38.858344 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:12:38 crc kubenswrapper[4742]: I1004 03:12:38.858366 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:12:38 crc kubenswrapper[4742]: I1004 03:12:38.858318 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:12:38 crc kubenswrapper[4742]: E1004 03:12:38.858466 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:12:38 crc kubenswrapper[4742]: E1004 03:12:38.858575 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:12:38 crc kubenswrapper[4742]: E1004 03:12:38.858614 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:12:38 crc kubenswrapper[4742]: E1004 03:12:38.858689 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:12:39 crc kubenswrapper[4742]: E1004 03:12:39.965890 4742 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 03:12:40 crc kubenswrapper[4742]: I1004 03:12:40.857670 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:12:40 crc kubenswrapper[4742]: I1004 03:12:40.857725 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:12:40 crc kubenswrapper[4742]: I1004 03:12:40.857664 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:12:40 crc kubenswrapper[4742]: I1004 03:12:40.857670 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:12:40 crc kubenswrapper[4742]: E1004 03:12:40.857838 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:12:40 crc kubenswrapper[4742]: E1004 03:12:40.857975 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:12:40 crc kubenswrapper[4742]: E1004 03:12:40.858065 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:12:40 crc kubenswrapper[4742]: E1004 03:12:40.858101 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:12:42 crc kubenswrapper[4742]: I1004 03:12:42.858322 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:12:42 crc kubenswrapper[4742]: I1004 03:12:42.858322 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:12:42 crc kubenswrapper[4742]: E1004 03:12:42.858495 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:12:42 crc kubenswrapper[4742]: I1004 03:12:42.858518 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:12:42 crc kubenswrapper[4742]: I1004 03:12:42.858594 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:12:42 crc kubenswrapper[4742]: E1004 03:12:42.858602 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:12:42 crc kubenswrapper[4742]: E1004 03:12:42.858749 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:12:42 crc kubenswrapper[4742]: E1004 03:12:42.858782 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:12:44 crc kubenswrapper[4742]: I1004 03:12:44.857721 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:12:44 crc kubenswrapper[4742]: I1004 03:12:44.857721 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:12:44 crc kubenswrapper[4742]: I1004 03:12:44.858364 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:12:44 crc kubenswrapper[4742]: E1004 03:12:44.859738 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:12:44 crc kubenswrapper[4742]: I1004 03:12:44.859823 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:12:44 crc kubenswrapper[4742]: E1004 03:12:44.859855 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:12:44 crc kubenswrapper[4742]: E1004 03:12:44.859964 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:12:44 crc kubenswrapper[4742]: E1004 03:12:44.860437 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:12:44 crc kubenswrapper[4742]: I1004 03:12:44.861914 4742 scope.go:117] "RemoveContainer" containerID="bfe5cd4a2153e5ae33beedc1cfdee7c39f57d2dc41a3d5e7d442df75d3c069db" Oct 04 03:12:44 crc kubenswrapper[4742]: E1004 03:12:44.966749 4742 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 03:12:45 crc kubenswrapper[4742]: I1004 03:12:45.512127 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nc8m7_46de52ee-226c-4713-bd7b-961f8f43f635/kube-multus/1.log" Oct 04 03:12:45 crc kubenswrapper[4742]: I1004 03:12:45.512204 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nc8m7" event={"ID":"46de52ee-226c-4713-bd7b-961f8f43f635","Type":"ContainerStarted","Data":"7a1d02fda520e7052329a1d1c9e5516ad70f170819a7d5263271e87b5928ba06"} Oct 04 03:12:46 crc kubenswrapper[4742]: I1004 03:12:46.858226 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:12:46 crc kubenswrapper[4742]: I1004 03:12:46.858309 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:12:46 crc kubenswrapper[4742]: I1004 03:12:46.858361 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:12:46 crc kubenswrapper[4742]: I1004 03:12:46.858309 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:12:46 crc kubenswrapper[4742]: E1004 03:12:46.858538 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:12:46 crc kubenswrapper[4742]: E1004 03:12:46.858895 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:12:46 crc kubenswrapper[4742]: E1004 03:12:46.858891 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:12:46 crc kubenswrapper[4742]: E1004 03:12:46.859025 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:12:48 crc kubenswrapper[4742]: I1004 03:12:48.857914 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:12:48 crc kubenswrapper[4742]: I1004 03:12:48.857966 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:12:48 crc kubenswrapper[4742]: I1004 03:12:48.858017 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:12:48 crc kubenswrapper[4742]: E1004 03:12:48.858107 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:12:48 crc kubenswrapper[4742]: I1004 03:12:48.858137 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:12:48 crc kubenswrapper[4742]: E1004 03:12:48.858330 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:12:48 crc kubenswrapper[4742]: E1004 03:12:48.858385 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cd46x" podUID="0302950c-deb3-420a-b798-c99459e62b3f" Oct 04 03:12:48 crc kubenswrapper[4742]: E1004 03:12:48.858453 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:12:50 crc kubenswrapper[4742]: I1004 03:12:50.857722 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:12:50 crc kubenswrapper[4742]: I1004 03:12:50.857838 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:12:50 crc kubenswrapper[4742]: I1004 03:12:50.857838 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:12:50 crc kubenswrapper[4742]: I1004 03:12:50.857865 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:12:50 crc kubenswrapper[4742]: I1004 03:12:50.862371 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 04 03:12:50 crc kubenswrapper[4742]: I1004 03:12:50.862707 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 04 03:12:50 crc kubenswrapper[4742]: I1004 03:12:50.862735 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 04 03:12:50 crc kubenswrapper[4742]: I1004 03:12:50.862914 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 04 03:12:50 crc kubenswrapper[4742]: I1004 03:12:50.863020 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 04 03:12:50 crc kubenswrapper[4742]: I1004 03:12:50.862928 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 04 03:12:50 crc kubenswrapper[4742]: I1004 03:12:50.989740 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:12:52 crc kubenswrapper[4742]: I1004 03:12:52.978337 4742 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.018028 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-f5978"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.018646 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.020904 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-t7pnt"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.021204 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t7pnt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.021554 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-gszdm"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.022081 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.022146 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-gszdm" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.023212 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-fcc5d"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.023556 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-fcc5d" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.026008 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-88xxl"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.026321 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-88xxl" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.032511 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.033387 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.038013 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.038593 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.038938 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.039074 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.039536 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.039538 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.039655 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.039792 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.040655 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.041689 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.041936 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.041990 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.041847 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.042116 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.042341 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.042425 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.042497 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.042570 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.042575 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.042451 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.042758 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.042799 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.042856 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.042911 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.042975 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.043016 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.043761 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.044107 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.044212 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mspvb"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.044657 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.044926 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-pkdzv"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.045563 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pkdzv" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.047416 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-2pdbs"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.067631 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.067894 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.068776 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-2pdbs" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.069787 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-ctkcp"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.070486 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ctkcp" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.073185 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.073379 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.074512 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.074710 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.074827 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.075225 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.084789 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.084910 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.084978 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.085192 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.085984 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.086894 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mgh2f"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.087427 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-fcc5d"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.087512 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mgh2f" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.088351 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-h22jn"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.088791 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-h22jn" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.089342 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-td4wb\" (UniqueName: \"kubernetes.io/projected/e537cd22-13ae-4f86-a246-aa53e62e8e4d-kube-api-access-td4wb\") pod \"apiserver-76f77b778f-f5978\" (UID: \"e537cd22-13ae-4f86-a246-aa53e62e8e4d\") " pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.089375 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f16638c6-63a8-41e3-9f0c-64fccee4ce3b-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-88xxl\" (UID: \"f16638c6-63a8-41e3-9f0c-64fccee4ce3b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-88xxl" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.089410 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/690f72b6-3754-442a-b1c9-59ff8ed83c1a-client-ca\") pod \"controller-manager-879f6c89f-gszdm\" (UID: \"690f72b6-3754-442a-b1c9-59ff8ed83c1a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gszdm" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.089432 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f16638c6-63a8-41e3-9f0c-64fccee4ce3b-config\") pod \"openshift-apiserver-operator-796bbdcf4f-88xxl\" (UID: \"f16638c6-63a8-41e3-9f0c-64fccee4ce3b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-88xxl" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.089449 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjkld\" (UniqueName: \"kubernetes.io/projected/f16638c6-63a8-41e3-9f0c-64fccee4ce3b-kube-api-access-fjkld\") pod \"openshift-apiserver-operator-796bbdcf4f-88xxl\" (UID: \"f16638c6-63a8-41e3-9f0c-64fccee4ce3b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-88xxl" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.089466 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e537cd22-13ae-4f86-a246-aa53e62e8e4d-serving-cert\") pod \"apiserver-76f77b778f-f5978\" (UID: \"e537cd22-13ae-4f86-a246-aa53e62e8e4d\") " pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.089485 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0062ea87-e9b5-44f9-88c2-27495348485e-images\") pod \"machine-api-operator-5694c8668f-fcc5d\" (UID: \"0062ea87-e9b5-44f9-88c2-27495348485e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fcc5d" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.089509 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/938dc4e6-1064-4eab-a3e3-8e3096109655-config\") pod \"route-controller-manager-6576b87f9c-t7pnt\" (UID: \"938dc4e6-1064-4eab-a3e3-8e3096109655\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t7pnt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.089534 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/938dc4e6-1064-4eab-a3e3-8e3096109655-client-ca\") pod \"route-controller-manager-6576b87f9c-t7pnt\" (UID: \"938dc4e6-1064-4eab-a3e3-8e3096109655\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t7pnt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.089554 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e537cd22-13ae-4f86-a246-aa53e62e8e4d-etcd-serving-ca\") pod \"apiserver-76f77b778f-f5978\" (UID: \"e537cd22-13ae-4f86-a246-aa53e62e8e4d\") " pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.089573 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0062ea87-e9b5-44f9-88c2-27495348485e-config\") pod \"machine-api-operator-5694c8668f-fcc5d\" (UID: \"0062ea87-e9b5-44f9-88c2-27495348485e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fcc5d" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.089592 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e537cd22-13ae-4f86-a246-aa53e62e8e4d-image-import-ca\") pod \"apiserver-76f77b778f-f5978\" (UID: \"e537cd22-13ae-4f86-a246-aa53e62e8e4d\") " pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.089606 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e537cd22-13ae-4f86-a246-aa53e62e8e4d-node-pullsecrets\") pod \"apiserver-76f77b778f-f5978\" (UID: \"e537cd22-13ae-4f86-a246-aa53e62e8e4d\") " pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.089628 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8flcs\" (UniqueName: \"kubernetes.io/projected/690f72b6-3754-442a-b1c9-59ff8ed83c1a-kube-api-access-8flcs\") pod \"controller-manager-879f6c89f-gszdm\" (UID: \"690f72b6-3754-442a-b1c9-59ff8ed83c1a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gszdm" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.089652 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e537cd22-13ae-4f86-a246-aa53e62e8e4d-etcd-client\") pod \"apiserver-76f77b778f-f5978\" (UID: \"e537cd22-13ae-4f86-a246-aa53e62e8e4d\") " pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.089666 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e537cd22-13ae-4f86-a246-aa53e62e8e4d-trusted-ca-bundle\") pod \"apiserver-76f77b778f-f5978\" (UID: \"e537cd22-13ae-4f86-a246-aa53e62e8e4d\") " pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.089686 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e537cd22-13ae-4f86-a246-aa53e62e8e4d-encryption-config\") pod \"apiserver-76f77b778f-f5978\" (UID: \"e537cd22-13ae-4f86-a246-aa53e62e8e4d\") " pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.089709 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/690f72b6-3754-442a-b1c9-59ff8ed83c1a-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-gszdm\" (UID: \"690f72b6-3754-442a-b1c9-59ff8ed83c1a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gszdm" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.089725 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/690f72b6-3754-442a-b1c9-59ff8ed83c1a-serving-cert\") pod \"controller-manager-879f6c89f-gszdm\" (UID: \"690f72b6-3754-442a-b1c9-59ff8ed83c1a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gszdm" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.089742 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngndk\" (UniqueName: \"kubernetes.io/projected/938dc4e6-1064-4eab-a3e3-8e3096109655-kube-api-access-ngndk\") pod \"route-controller-manager-6576b87f9c-t7pnt\" (UID: \"938dc4e6-1064-4eab-a3e3-8e3096109655\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t7pnt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.089765 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/690f72b6-3754-442a-b1c9-59ff8ed83c1a-config\") pod \"controller-manager-879f6c89f-gszdm\" (UID: \"690f72b6-3754-442a-b1c9-59ff8ed83c1a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gszdm" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.089787 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/0062ea87-e9b5-44f9-88c2-27495348485e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-fcc5d\" (UID: \"0062ea87-e9b5-44f9-88c2-27495348485e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fcc5d" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.089812 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/938dc4e6-1064-4eab-a3e3-8e3096109655-serving-cert\") pod \"route-controller-manager-6576b87f9c-t7pnt\" (UID: \"938dc4e6-1064-4eab-a3e3-8e3096109655\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t7pnt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.089832 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e537cd22-13ae-4f86-a246-aa53e62e8e4d-audit-dir\") pod \"apiserver-76f77b778f-f5978\" (UID: \"e537cd22-13ae-4f86-a246-aa53e62e8e4d\") " pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.089850 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6w26\" (UniqueName: \"kubernetes.io/projected/0062ea87-e9b5-44f9-88c2-27495348485e-kube-api-access-b6w26\") pod \"machine-api-operator-5694c8668f-fcc5d\" (UID: \"0062ea87-e9b5-44f9-88c2-27495348485e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fcc5d" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.089867 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e537cd22-13ae-4f86-a246-aa53e62e8e4d-config\") pod \"apiserver-76f77b778f-f5978\" (UID: \"e537cd22-13ae-4f86-a246-aa53e62e8e4d\") " pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.089884 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e537cd22-13ae-4f86-a246-aa53e62e8e4d-audit\") pod \"apiserver-76f77b778f-f5978\" (UID: \"e537cd22-13ae-4f86-a246-aa53e62e8e4d\") " pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.090025 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.090500 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.091556 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6nmn2"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.091600 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.091708 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.091962 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.092083 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6nmn2" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.092225 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.092477 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.092593 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.094256 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.094988 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.095450 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.095492 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.095578 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.095792 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.096313 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.096884 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rlbvb"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.097185 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.097471 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rlbvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.097522 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.097724 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-8w75v"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.098428 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8w75v" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.098570 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.099738 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.100386 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.100925 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-cms9s"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.101435 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.101573 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-cms9s" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.101835 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-v92dt"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.102152 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.102376 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-v92dt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.104296 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-mfp6b"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.106604 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nsspd"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.104835 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.121020 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mfp6b" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.124012 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-dmd8c"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.125144 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-dmd8c" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.105067 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.125875 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.127223 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.129585 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-fvr9s"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.134382 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-f6gcs"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.135245 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-fvr9s" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.137514 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nsspd" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.140367 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ncmp2"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.141585 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.142601 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ncmp2" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.159045 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.159394 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.159605 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.160887 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.162637 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.163808 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.164436 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.164532 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.164618 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.164715 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.164832 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-p559m"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.164858 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.165096 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.165376 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.165512 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.165537 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.165428 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.165474 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.165544 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-p559m" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.165680 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.165767 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.165447 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.165900 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.166006 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.166047 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.166110 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.166158 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.166168 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.166205 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.166009 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.166290 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.166021 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.166296 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.166412 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.166577 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.166623 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.169310 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.171128 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-88xxl"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.171194 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-t7pnt"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.171591 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-85fhx"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.172184 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-85fhx" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.174340 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.176535 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.185750 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.190582 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/690f72b6-3754-442a-b1c9-59ff8ed83c1a-config\") pod \"controller-manager-879f6c89f-gszdm\" (UID: \"690f72b6-3754-442a-b1c9-59ff8ed83c1a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gszdm" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.190623 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/0062ea87-e9b5-44f9-88c2-27495348485e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-fcc5d\" (UID: \"0062ea87-e9b5-44f9-88c2-27495348485e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fcc5d" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.190648 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqbln\" (UniqueName: \"kubernetes.io/projected/da95bbc1-a1c0-428f-9884-cda886a8da3c-kube-api-access-rqbln\") pod \"machine-approver-56656f9798-pkdzv\" (UID: \"da95bbc1-a1c0-428f-9884-cda886a8da3c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pkdzv" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.190666 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09503673-286e-48af-b18b-37bff87f8019-config\") pod \"authentication-operator-69f744f599-fvr9s\" (UID: \"09503673-286e-48af-b18b-37bff87f8019\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fvr9s" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.190690 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/938dc4e6-1064-4eab-a3e3-8e3096109655-serving-cert\") pod \"route-controller-manager-6576b87f9c-t7pnt\" (UID: \"938dc4e6-1064-4eab-a3e3-8e3096109655\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t7pnt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.190709 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e537cd22-13ae-4f86-a246-aa53e62e8e4d-audit-dir\") pod \"apiserver-76f77b778f-f5978\" (UID: \"e537cd22-13ae-4f86-a246-aa53e62e8e4d\") " pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.190727 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/56722804-fa7c-4639-a4d3-5085cd965f8b-audit-policies\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.190742 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/da95bbc1-a1c0-428f-9884-cda886a8da3c-auth-proxy-config\") pod \"machine-approver-56656f9798-pkdzv\" (UID: \"da95bbc1-a1c0-428f-9884-cda886a8da3c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pkdzv" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.190761 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6w26\" (UniqueName: \"kubernetes.io/projected/0062ea87-e9b5-44f9-88c2-27495348485e-kube-api-access-b6w26\") pod \"machine-api-operator-5694c8668f-fcc5d\" (UID: \"0062ea87-e9b5-44f9-88c2-27495348485e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fcc5d" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.190779 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.190798 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8725e189-8b15-4c13-ae9b-f1230476e572-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-nsspd\" (UID: \"8725e189-8b15-4c13-ae9b-f1230476e572\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nsspd" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.190816 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/56722804-fa7c-4639-a4d3-5085cd965f8b-audit-dir\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.190835 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e537cd22-13ae-4f86-a246-aa53e62e8e4d-config\") pod \"apiserver-76f77b778f-f5978\" (UID: \"e537cd22-13ae-4f86-a246-aa53e62e8e4d\") " pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.190850 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e537cd22-13ae-4f86-a246-aa53e62e8e4d-audit\") pod \"apiserver-76f77b778f-f5978\" (UID: \"e537cd22-13ae-4f86-a246-aa53e62e8e4d\") " pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.190868 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.190886 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e3b81c30-6a85-430c-a915-6322d20953ac-metrics-tls\") pod \"ingress-operator-5b745b69d9-mfp6b\" (UID: \"e3b81c30-6a85-430c-a915-6322d20953ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mfp6b" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.190903 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07f1cf4a-506e-469a-a8e5-0adc68a5ab81-config\") pod \"console-operator-58897d9998-2pdbs\" (UID: \"07f1cf4a-506e-469a-a8e5-0adc68a5ab81\") " pod="openshift-console-operator/console-operator-58897d9998-2pdbs" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.190923 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ls9r\" (UniqueName: \"kubernetes.io/projected/07f1cf4a-506e-469a-a8e5-0adc68a5ab81-kube-api-access-2ls9r\") pod \"console-operator-58897d9998-2pdbs\" (UID: \"07f1cf4a-506e-469a-a8e5-0adc68a5ab81\") " pod="openshift-console-operator/console-operator-58897d9998-2pdbs" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.190939 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5ba77307-01a4-43a3-8888-3da1c700ea2e-oauth-serving-cert\") pod \"console-f9d7485db-ctkcp\" (UID: \"5ba77307-01a4-43a3-8888-3da1c700ea2e\") " pod="openshift-console/console-f9d7485db-ctkcp" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.190955 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09503673-286e-48af-b18b-37bff87f8019-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-fvr9s\" (UID: \"09503673-286e-48af-b18b-37bff87f8019\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fvr9s" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.190971 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5ba77307-01a4-43a3-8888-3da1c700ea2e-service-ca\") pod \"console-f9d7485db-ctkcp\" (UID: \"5ba77307-01a4-43a3-8888-3da1c700ea2e\") " pod="openshift-console/console-f9d7485db-ctkcp" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.190988 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-td4wb\" (UniqueName: \"kubernetes.io/projected/e537cd22-13ae-4f86-a246-aa53e62e8e4d-kube-api-access-td4wb\") pod \"apiserver-76f77b778f-f5978\" (UID: \"e537cd22-13ae-4f86-a246-aa53e62e8e4d\") " pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191019 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f16638c6-63a8-41e3-9f0c-64fccee4ce3b-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-88xxl\" (UID: \"f16638c6-63a8-41e3-9f0c-64fccee4ce3b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-88xxl" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191042 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9w5jf\" (UniqueName: \"kubernetes.io/projected/56722804-fa7c-4639-a4d3-5085cd965f8b-kube-api-access-9w5jf\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191066 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e3b81c30-6a85-430c-a915-6322d20953ac-bound-sa-token\") pod \"ingress-operator-5b745b69d9-mfp6b\" (UID: \"e3b81c30-6a85-430c-a915-6322d20953ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mfp6b" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191088 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/07f1cf4a-506e-469a-a8e5-0adc68a5ab81-trusted-ca\") pod \"console-operator-58897d9998-2pdbs\" (UID: \"07f1cf4a-506e-469a-a8e5-0adc68a5ab81\") " pod="openshift-console-operator/console-operator-58897d9998-2pdbs" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191126 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/690f72b6-3754-442a-b1c9-59ff8ed83c1a-client-ca\") pod \"controller-manager-879f6c89f-gszdm\" (UID: \"690f72b6-3754-442a-b1c9-59ff8ed83c1a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gszdm" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191147 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f16638c6-63a8-41e3-9f0c-64fccee4ce3b-config\") pod \"openshift-apiserver-operator-796bbdcf4f-88xxl\" (UID: \"f16638c6-63a8-41e3-9f0c-64fccee4ce3b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-88xxl" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191168 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e3b81c30-6a85-430c-a915-6322d20953ac-trusted-ca\") pod \"ingress-operator-5b745b69d9-mfp6b\" (UID: \"e3b81c30-6a85-430c-a915-6322d20953ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mfp6b" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191194 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjkld\" (UniqueName: \"kubernetes.io/projected/f16638c6-63a8-41e3-9f0c-64fccee4ce3b-kube-api-access-fjkld\") pod \"openshift-apiserver-operator-796bbdcf4f-88xxl\" (UID: \"f16638c6-63a8-41e3-9f0c-64fccee4ce3b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-88xxl" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191219 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09503673-286e-48af-b18b-37bff87f8019-serving-cert\") pod \"authentication-operator-69f744f599-fvr9s\" (UID: \"09503673-286e-48af-b18b-37bff87f8019\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fvr9s" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191245 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e537cd22-13ae-4f86-a246-aa53e62e8e4d-serving-cert\") pod \"apiserver-76f77b778f-f5978\" (UID: \"e537cd22-13ae-4f86-a246-aa53e62e8e4d\") " pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191293 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191320 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5ba77307-01a4-43a3-8888-3da1c700ea2e-console-serving-cert\") pod \"console-f9d7485db-ctkcp\" (UID: \"5ba77307-01a4-43a3-8888-3da1c700ea2e\") " pod="openshift-console/console-f9d7485db-ctkcp" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191348 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191373 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191396 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkrcg\" (UniqueName: \"kubernetes.io/projected/0e5511f2-7587-4638-bd0c-af91c57a0163-kube-api-access-dkrcg\") pod \"downloads-7954f5f757-h22jn\" (UID: \"0e5511f2-7587-4638-bd0c-af91c57a0163\") " pod="openshift-console/downloads-7954f5f757-h22jn" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191424 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/938dc4e6-1064-4eab-a3e3-8e3096109655-config\") pod \"route-controller-manager-6576b87f9c-t7pnt\" (UID: \"938dc4e6-1064-4eab-a3e3-8e3096109655\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t7pnt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191449 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/938dc4e6-1064-4eab-a3e3-8e3096109655-client-ca\") pod \"route-controller-manager-6576b87f9c-t7pnt\" (UID: \"938dc4e6-1064-4eab-a3e3-8e3096109655\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t7pnt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191470 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e537cd22-13ae-4f86-a246-aa53e62e8e4d-etcd-serving-ca\") pod \"apiserver-76f77b778f-f5978\" (UID: \"e537cd22-13ae-4f86-a246-aa53e62e8e4d\") " pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191489 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0062ea87-e9b5-44f9-88c2-27495348485e-config\") pod \"machine-api-operator-5694c8668f-fcc5d\" (UID: \"0062ea87-e9b5-44f9-88c2-27495348485e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fcc5d" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191507 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0062ea87-e9b5-44f9-88c2-27495348485e-images\") pod \"machine-api-operator-5694c8668f-fcc5d\" (UID: \"0062ea87-e9b5-44f9-88c2-27495348485e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fcc5d" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191527 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191550 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvq2q\" (UniqueName: \"kubernetes.io/projected/5ba77307-01a4-43a3-8888-3da1c700ea2e-kube-api-access-qvq2q\") pod \"console-f9d7485db-ctkcp\" (UID: \"5ba77307-01a4-43a3-8888-3da1c700ea2e\") " pod="openshift-console/console-f9d7485db-ctkcp" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191567 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e537cd22-13ae-4f86-a246-aa53e62e8e4d-image-import-ca\") pod \"apiserver-76f77b778f-f5978\" (UID: \"e537cd22-13ae-4f86-a246-aa53e62e8e4d\") " pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191584 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8725e189-8b15-4c13-ae9b-f1230476e572-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-nsspd\" (UID: \"8725e189-8b15-4c13-ae9b-f1230476e572\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nsspd" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191601 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgfbg\" (UniqueName: \"kubernetes.io/projected/e3b81c30-6a85-430c-a915-6322d20953ac-kube-api-access-wgfbg\") pod \"ingress-operator-5b745b69d9-mfp6b\" (UID: \"e3b81c30-6a85-430c-a915-6322d20953ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mfp6b" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191620 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5ba77307-01a4-43a3-8888-3da1c700ea2e-console-oauth-config\") pod \"console-f9d7485db-ctkcp\" (UID: \"5ba77307-01a4-43a3-8888-3da1c700ea2e\") " pod="openshift-console/console-f9d7485db-ctkcp" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191637 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e537cd22-13ae-4f86-a246-aa53e62e8e4d-node-pullsecrets\") pod \"apiserver-76f77b778f-f5978\" (UID: \"e537cd22-13ae-4f86-a246-aa53e62e8e4d\") " pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191660 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8flcs\" (UniqueName: \"kubernetes.io/projected/690f72b6-3754-442a-b1c9-59ff8ed83c1a-kube-api-access-8flcs\") pod \"controller-manager-879f6c89f-gszdm\" (UID: \"690f72b6-3754-442a-b1c9-59ff8ed83c1a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gszdm" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191682 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/c975f343-d9c0-4145-b0b8-bdc11dbd1d66-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-mgh2f\" (UID: \"c975f343-d9c0-4145-b0b8-bdc11dbd1d66\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mgh2f" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191701 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c975f343-d9c0-4145-b0b8-bdc11dbd1d66-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-mgh2f\" (UID: \"c975f343-d9c0-4145-b0b8-bdc11dbd1d66\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mgh2f" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191724 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c975f343-d9c0-4145-b0b8-bdc11dbd1d66-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-mgh2f\" (UID: \"c975f343-d9c0-4145-b0b8-bdc11dbd1d66\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mgh2f" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191760 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da95bbc1-a1c0-428f-9884-cda886a8da3c-config\") pod \"machine-approver-56656f9798-pkdzv\" (UID: \"da95bbc1-a1c0-428f-9884-cda886a8da3c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pkdzv" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191788 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6c8p8\" (UniqueName: \"kubernetes.io/projected/c975f343-d9c0-4145-b0b8-bdc11dbd1d66-kube-api-access-6c8p8\") pod \"cluster-image-registry-operator-dc59b4c8b-mgh2f\" (UID: \"c975f343-d9c0-4145-b0b8-bdc11dbd1d66\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mgh2f" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191817 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pd8s\" (UniqueName: \"kubernetes.io/projected/8725e189-8b15-4c13-ae9b-f1230476e572-kube-api-access-9pd8s\") pod \"kube-storage-version-migrator-operator-b67b599dd-nsspd\" (UID: \"8725e189-8b15-4c13-ae9b-f1230476e572\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nsspd" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191833 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07f1cf4a-506e-469a-a8e5-0adc68a5ab81-serving-cert\") pod \"console-operator-58897d9998-2pdbs\" (UID: \"07f1cf4a-506e-469a-a8e5-0adc68a5ab81\") " pod="openshift-console-operator/console-operator-58897d9998-2pdbs" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191850 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e537cd22-13ae-4f86-a246-aa53e62e8e4d-etcd-client\") pod \"apiserver-76f77b778f-f5978\" (UID: \"e537cd22-13ae-4f86-a246-aa53e62e8e4d\") " pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191866 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e537cd22-13ae-4f86-a246-aa53e62e8e4d-trusted-ca-bundle\") pod \"apiserver-76f77b778f-f5978\" (UID: \"e537cd22-13ae-4f86-a246-aa53e62e8e4d\") " pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191886 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e537cd22-13ae-4f86-a246-aa53e62e8e4d-encryption-config\") pod \"apiserver-76f77b778f-f5978\" (UID: \"e537cd22-13ae-4f86-a246-aa53e62e8e4d\") " pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191904 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191922 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191948 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191973 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.191992 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.192009 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5ba77307-01a4-43a3-8888-3da1c700ea2e-console-config\") pod \"console-f9d7485db-ctkcp\" (UID: \"5ba77307-01a4-43a3-8888-3da1c700ea2e\") " pod="openshift-console/console-f9d7485db-ctkcp" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.192037 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/690f72b6-3754-442a-b1c9-59ff8ed83c1a-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-gszdm\" (UID: \"690f72b6-3754-442a-b1c9-59ff8ed83c1a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gszdm" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.192057 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/690f72b6-3754-442a-b1c9-59ff8ed83c1a-serving-cert\") pod \"controller-manager-879f6c89f-gszdm\" (UID: \"690f72b6-3754-442a-b1c9-59ff8ed83c1a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gszdm" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.192075 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngndk\" (UniqueName: \"kubernetes.io/projected/938dc4e6-1064-4eab-a3e3-8e3096109655-kube-api-access-ngndk\") pod \"route-controller-manager-6576b87f9c-t7pnt\" (UID: \"938dc4e6-1064-4eab-a3e3-8e3096109655\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t7pnt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.192091 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/da95bbc1-a1c0-428f-9884-cda886a8da3c-machine-approver-tls\") pod \"machine-approver-56656f9798-pkdzv\" (UID: \"da95bbc1-a1c0-428f-9884-cda886a8da3c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pkdzv" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.192110 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09503673-286e-48af-b18b-37bff87f8019-service-ca-bundle\") pod \"authentication-operator-69f744f599-fvr9s\" (UID: \"09503673-286e-48af-b18b-37bff87f8019\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fvr9s" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.192127 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rplt\" (UniqueName: \"kubernetes.io/projected/09503673-286e-48af-b18b-37bff87f8019-kube-api-access-2rplt\") pod \"authentication-operator-69f744f599-fvr9s\" (UID: \"09503673-286e-48af-b18b-37bff87f8019\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fvr9s" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.192145 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5ba77307-01a4-43a3-8888-3da1c700ea2e-trusted-ca-bundle\") pod \"console-f9d7485db-ctkcp\" (UID: \"5ba77307-01a4-43a3-8888-3da1c700ea2e\") " pod="openshift-console/console-f9d7485db-ctkcp" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.194263 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/938dc4e6-1064-4eab-a3e3-8e3096109655-config\") pod \"route-controller-manager-6576b87f9c-t7pnt\" (UID: \"938dc4e6-1064-4eab-a3e3-8e3096109655\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t7pnt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.195661 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e537cd22-13ae-4f86-a246-aa53e62e8e4d-audit-dir\") pod \"apiserver-76f77b778f-f5978\" (UID: \"e537cd22-13ae-4f86-a246-aa53e62e8e4d\") " pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.196555 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e537cd22-13ae-4f86-a246-aa53e62e8e4d-trusted-ca-bundle\") pod \"apiserver-76f77b778f-f5978\" (UID: \"e537cd22-13ae-4f86-a246-aa53e62e8e4d\") " pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.198022 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/938dc4e6-1064-4eab-a3e3-8e3096109655-client-ca\") pod \"route-controller-manager-6576b87f9c-t7pnt\" (UID: \"938dc4e6-1064-4eab-a3e3-8e3096109655\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t7pnt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.198457 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e537cd22-13ae-4f86-a246-aa53e62e8e4d-etcd-serving-ca\") pod \"apiserver-76f77b778f-f5978\" (UID: \"e537cd22-13ae-4f86-a246-aa53e62e8e4d\") " pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.199121 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0062ea87-e9b5-44f9-88c2-27495348485e-config\") pod \"machine-api-operator-5694c8668f-fcc5d\" (UID: \"0062ea87-e9b5-44f9-88c2-27495348485e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fcc5d" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.199553 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f16638c6-63a8-41e3-9f0c-64fccee4ce3b-config\") pod \"openshift-apiserver-operator-796bbdcf4f-88xxl\" (UID: \"f16638c6-63a8-41e3-9f0c-64fccee4ce3b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-88xxl" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.199700 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0062ea87-e9b5-44f9-88c2-27495348485e-images\") pod \"machine-api-operator-5694c8668f-fcc5d\" (UID: \"0062ea87-e9b5-44f9-88c2-27495348485e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fcc5d" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.203302 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e537cd22-13ae-4f86-a246-aa53e62e8e4d-encryption-config\") pod \"apiserver-76f77b778f-f5978\" (UID: \"e537cd22-13ae-4f86-a246-aa53e62e8e4d\") " pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.203883 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e537cd22-13ae-4f86-a246-aa53e62e8e4d-node-pullsecrets\") pod \"apiserver-76f77b778f-f5978\" (UID: \"e537cd22-13ae-4f86-a246-aa53e62e8e4d\") " pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.204355 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kfp77"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.204533 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e537cd22-13ae-4f86-a246-aa53e62e8e4d-audit\") pod \"apiserver-76f77b778f-f5978\" (UID: \"e537cd22-13ae-4f86-a246-aa53e62e8e4d\") " pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.204996 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tfc7g"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.205156 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e537cd22-13ae-4f86-a246-aa53e62e8e4d-config\") pod \"apiserver-76f77b778f-f5978\" (UID: \"e537cd22-13ae-4f86-a246-aa53e62e8e4d\") " pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.205400 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/690f72b6-3754-442a-b1c9-59ff8ed83c1a-client-ca\") pod \"controller-manager-879f6c89f-gszdm\" (UID: \"690f72b6-3754-442a-b1c9-59ff8ed83c1a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gszdm" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.206010 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e537cd22-13ae-4f86-a246-aa53e62e8e4d-image-import-ca\") pod \"apiserver-76f77b778f-f5978\" (UID: \"e537cd22-13ae-4f86-a246-aa53e62e8e4d\") " pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.206084 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kfp77" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.206398 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-tfc7g" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.206604 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/690f72b6-3754-442a-b1c9-59ff8ed83c1a-config\") pod \"controller-manager-879f6c89f-gszdm\" (UID: \"690f72b6-3754-442a-b1c9-59ff8ed83c1a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gszdm" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.206752 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/690f72b6-3754-442a-b1c9-59ff8ed83c1a-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-gszdm\" (UID: \"690f72b6-3754-442a-b1c9-59ff8ed83c1a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gszdm" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.209575 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/690f72b6-3754-442a-b1c9-59ff8ed83c1a-serving-cert\") pod \"controller-manager-879f6c89f-gszdm\" (UID: \"690f72b6-3754-442a-b1c9-59ff8ed83c1a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gszdm" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.209642 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rtd5w"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.210318 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-6prnz"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.211061 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6prnz" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.211454 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rtd5w" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.211513 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325780-9b22t"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.211974 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325780-9b22t" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.213660 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-f5978"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.214097 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f16638c6-63a8-41e3-9f0c-64fccee4ce3b-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-88xxl\" (UID: \"f16638c6-63a8-41e3-9f0c-64fccee4ce3b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-88xxl" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.215034 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/0062ea87-e9b5-44f9-88c2-27495348485e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-fcc5d\" (UID: \"0062ea87-e9b5-44f9-88c2-27495348485e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fcc5d" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.215086 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-4v795"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.215460 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-4v795" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.216217 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qznqm"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.216968 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qznqm" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.217095 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e537cd22-13ae-4f86-a246-aa53e62e8e4d-etcd-client\") pod \"apiserver-76f77b778f-f5978\" (UID: \"e537cd22-13ae-4f86-a246-aa53e62e8e4d\") " pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.217405 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e537cd22-13ae-4f86-a246-aa53e62e8e4d-serving-cert\") pod \"apiserver-76f77b778f-f5978\" (UID: \"e537cd22-13ae-4f86-a246-aa53e62e8e4d\") " pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.219453 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5j88f"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.220003 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5j88f" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.222335 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/938dc4e6-1064-4eab-a3e3-8e3096109655-serving-cert\") pod \"route-controller-manager-6576b87f9c-t7pnt\" (UID: \"938dc4e6-1064-4eab-a3e3-8e3096109655\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t7pnt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.222468 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.223366 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8kfz8"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.223988 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8kfz8" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.224349 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.225718 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xg5l9"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.226446 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xg5l9" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.227642 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-ktnrj"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.228187 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-ktnrj" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.230399 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-krpzs"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.231220 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-krpzs" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.232452 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-k77jn"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.236217 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-k77jn" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.237122 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-vtmjj"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.238672 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-vtmjj" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.242684 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mspvb"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.243742 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-lgfzr"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.243902 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.245400 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-lgfzr" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.245855 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-h22jn"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.249108 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rlbvb"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.250839 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-cms9s"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.254625 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6nmn2"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.255807 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-ctkcp"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.256925 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.261081 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.263683 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-mfp6b"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.268369 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-f6gcs"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.271050 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-2pdbs"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.274161 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-6prnz"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.275119 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-ktnrj"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.276319 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-8w75v"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.277926 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kfp77"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.278997 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-gszdm"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.280224 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-v92dt"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.281111 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.281628 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-dmd8c"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.282915 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nsspd"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.283916 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-wn8jl"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.285491 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-nvt5z"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.285607 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-wn8jl" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.286590 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-p559m"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.286806 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-nvt5z" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.287711 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-fvr9s"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.289054 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ncmp2"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.289961 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mgh2f"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.291264 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-k77jn"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.292258 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qznqm"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.292706 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09503673-286e-48af-b18b-37bff87f8019-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-fvr9s\" (UID: \"09503673-286e-48af-b18b-37bff87f8019\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fvr9s" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.292736 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5ba77307-01a4-43a3-8888-3da1c700ea2e-service-ca\") pod \"console-f9d7485db-ctkcp\" (UID: \"5ba77307-01a4-43a3-8888-3da1c700ea2e\") " pod="openshift-console/console-f9d7485db-ctkcp" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.292757 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9w5jf\" (UniqueName: \"kubernetes.io/projected/56722804-fa7c-4639-a4d3-5085cd965f8b-kube-api-access-9w5jf\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.292773 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e3b81c30-6a85-430c-a915-6322d20953ac-bound-sa-token\") pod \"ingress-operator-5b745b69d9-mfp6b\" (UID: \"e3b81c30-6a85-430c-a915-6322d20953ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mfp6b" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.292790 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/07f1cf4a-506e-469a-a8e5-0adc68a5ab81-trusted-ca\") pod \"console-operator-58897d9998-2pdbs\" (UID: \"07f1cf4a-506e-469a-a8e5-0adc68a5ab81\") " pod="openshift-console-operator/console-operator-58897d9998-2pdbs" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.292813 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e3b81c30-6a85-430c-a915-6322d20953ac-trusted-ca\") pod \"ingress-operator-5b745b69d9-mfp6b\" (UID: \"e3b81c30-6a85-430c-a915-6322d20953ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mfp6b" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.292845 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09503673-286e-48af-b18b-37bff87f8019-serving-cert\") pod \"authentication-operator-69f744f599-fvr9s\" (UID: \"09503673-286e-48af-b18b-37bff87f8019\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fvr9s" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.292863 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.292878 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5ba77307-01a4-43a3-8888-3da1c700ea2e-console-serving-cert\") pod \"console-f9d7485db-ctkcp\" (UID: \"5ba77307-01a4-43a3-8888-3da1c700ea2e\") " pod="openshift-console/console-f9d7485db-ctkcp" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.292899 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.292914 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.292928 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkrcg\" (UniqueName: \"kubernetes.io/projected/0e5511f2-7587-4638-bd0c-af91c57a0163-kube-api-access-dkrcg\") pod \"downloads-7954f5f757-h22jn\" (UID: \"0e5511f2-7587-4638-bd0c-af91c57a0163\") " pod="openshift-console/downloads-7954f5f757-h22jn" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.292947 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.292969 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvq2q\" (UniqueName: \"kubernetes.io/projected/5ba77307-01a4-43a3-8888-3da1c700ea2e-kube-api-access-qvq2q\") pod \"console-f9d7485db-ctkcp\" (UID: \"5ba77307-01a4-43a3-8888-3da1c700ea2e\") " pod="openshift-console/console-f9d7485db-ctkcp" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.292986 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8725e189-8b15-4c13-ae9b-f1230476e572-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-nsspd\" (UID: \"8725e189-8b15-4c13-ae9b-f1230476e572\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nsspd" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.293000 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgfbg\" (UniqueName: \"kubernetes.io/projected/e3b81c30-6a85-430c-a915-6322d20953ac-kube-api-access-wgfbg\") pod \"ingress-operator-5b745b69d9-mfp6b\" (UID: \"e3b81c30-6a85-430c-a915-6322d20953ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mfp6b" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.293017 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5ba77307-01a4-43a3-8888-3da1c700ea2e-console-oauth-config\") pod \"console-f9d7485db-ctkcp\" (UID: \"5ba77307-01a4-43a3-8888-3da1c700ea2e\") " pod="openshift-console/console-f9d7485db-ctkcp" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.293036 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/c975f343-d9c0-4145-b0b8-bdc11dbd1d66-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-mgh2f\" (UID: \"c975f343-d9c0-4145-b0b8-bdc11dbd1d66\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mgh2f" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.293052 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c975f343-d9c0-4145-b0b8-bdc11dbd1d66-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-mgh2f\" (UID: \"c975f343-d9c0-4145-b0b8-bdc11dbd1d66\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mgh2f" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.293068 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c975f343-d9c0-4145-b0b8-bdc11dbd1d66-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-mgh2f\" (UID: \"c975f343-d9c0-4145-b0b8-bdc11dbd1d66\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mgh2f" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.293101 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da95bbc1-a1c0-428f-9884-cda886a8da3c-config\") pod \"machine-approver-56656f9798-pkdzv\" (UID: \"da95bbc1-a1c0-428f-9884-cda886a8da3c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pkdzv" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.293122 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6c8p8\" (UniqueName: \"kubernetes.io/projected/c975f343-d9c0-4145-b0b8-bdc11dbd1d66-kube-api-access-6c8p8\") pod \"cluster-image-registry-operator-dc59b4c8b-mgh2f\" (UID: \"c975f343-d9c0-4145-b0b8-bdc11dbd1d66\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mgh2f" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.293152 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pd8s\" (UniqueName: \"kubernetes.io/projected/8725e189-8b15-4c13-ae9b-f1230476e572-kube-api-access-9pd8s\") pod \"kube-storage-version-migrator-operator-b67b599dd-nsspd\" (UID: \"8725e189-8b15-4c13-ae9b-f1230476e572\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nsspd" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.293202 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07f1cf4a-506e-469a-a8e5-0adc68a5ab81-serving-cert\") pod \"console-operator-58897d9998-2pdbs\" (UID: \"07f1cf4a-506e-469a-a8e5-0adc68a5ab81\") " pod="openshift-console-operator/console-operator-58897d9998-2pdbs" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.293225 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.293264 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.293304 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.293324 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.293342 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.293360 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5ba77307-01a4-43a3-8888-3da1c700ea2e-console-config\") pod \"console-f9d7485db-ctkcp\" (UID: \"5ba77307-01a4-43a3-8888-3da1c700ea2e\") " pod="openshift-console/console-f9d7485db-ctkcp" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.293388 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/da95bbc1-a1c0-428f-9884-cda886a8da3c-machine-approver-tls\") pod \"machine-approver-56656f9798-pkdzv\" (UID: \"da95bbc1-a1c0-428f-9884-cda886a8da3c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pkdzv" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.293410 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09503673-286e-48af-b18b-37bff87f8019-service-ca-bundle\") pod \"authentication-operator-69f744f599-fvr9s\" (UID: \"09503673-286e-48af-b18b-37bff87f8019\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fvr9s" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.293428 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rplt\" (UniqueName: \"kubernetes.io/projected/09503673-286e-48af-b18b-37bff87f8019-kube-api-access-2rplt\") pod \"authentication-operator-69f744f599-fvr9s\" (UID: \"09503673-286e-48af-b18b-37bff87f8019\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fvr9s" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.293444 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5ba77307-01a4-43a3-8888-3da1c700ea2e-trusted-ca-bundle\") pod \"console-f9d7485db-ctkcp\" (UID: \"5ba77307-01a4-43a3-8888-3da1c700ea2e\") " pod="openshift-console/console-f9d7485db-ctkcp" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.293465 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqbln\" (UniqueName: \"kubernetes.io/projected/da95bbc1-a1c0-428f-9884-cda886a8da3c-kube-api-access-rqbln\") pod \"machine-approver-56656f9798-pkdzv\" (UID: \"da95bbc1-a1c0-428f-9884-cda886a8da3c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pkdzv" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.293483 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09503673-286e-48af-b18b-37bff87f8019-config\") pod \"authentication-operator-69f744f599-fvr9s\" (UID: \"09503673-286e-48af-b18b-37bff87f8019\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fvr9s" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.293502 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/56722804-fa7c-4639-a4d3-5085cd965f8b-audit-policies\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.293519 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/da95bbc1-a1c0-428f-9884-cda886a8da3c-auth-proxy-config\") pod \"machine-approver-56656f9798-pkdzv\" (UID: \"da95bbc1-a1c0-428f-9884-cda886a8da3c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pkdzv" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.293546 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.293567 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8725e189-8b15-4c13-ae9b-f1230476e572-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-nsspd\" (UID: \"8725e189-8b15-4c13-ae9b-f1230476e572\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nsspd" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.293589 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/56722804-fa7c-4639-a4d3-5085cd965f8b-audit-dir\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.293611 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.293627 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e3b81c30-6a85-430c-a915-6322d20953ac-metrics-tls\") pod \"ingress-operator-5b745b69d9-mfp6b\" (UID: \"e3b81c30-6a85-430c-a915-6322d20953ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mfp6b" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.293642 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07f1cf4a-506e-469a-a8e5-0adc68a5ab81-config\") pod \"console-operator-58897d9998-2pdbs\" (UID: \"07f1cf4a-506e-469a-a8e5-0adc68a5ab81\") " pod="openshift-console-operator/console-operator-58897d9998-2pdbs" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.293659 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5ba77307-01a4-43a3-8888-3da1c700ea2e-oauth-serving-cert\") pod \"console-f9d7485db-ctkcp\" (UID: \"5ba77307-01a4-43a3-8888-3da1c700ea2e\") " pod="openshift-console/console-f9d7485db-ctkcp" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.293675 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ls9r\" (UniqueName: \"kubernetes.io/projected/07f1cf4a-506e-469a-a8e5-0adc68a5ab81-kube-api-access-2ls9r\") pod \"console-operator-58897d9998-2pdbs\" (UID: \"07f1cf4a-506e-469a-a8e5-0adc68a5ab81\") " pod="openshift-console-operator/console-operator-58897d9998-2pdbs" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.293729 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-krpzs"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.294446 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.295145 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5ba77307-01a4-43a3-8888-3da1c700ea2e-service-ca\") pod \"console-f9d7485db-ctkcp\" (UID: \"5ba77307-01a4-43a3-8888-3da1c700ea2e\") " pod="openshift-console/console-f9d7485db-ctkcp" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.295831 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c975f343-d9c0-4145-b0b8-bdc11dbd1d66-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-mgh2f\" (UID: \"c975f343-d9c0-4145-b0b8-bdc11dbd1d66\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mgh2f" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.296262 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da95bbc1-a1c0-428f-9884-cda886a8da3c-config\") pod \"machine-approver-56656f9798-pkdzv\" (UID: \"da95bbc1-a1c0-428f-9884-cda886a8da3c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pkdzv" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.296360 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.297067 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/07f1cf4a-506e-469a-a8e5-0adc68a5ab81-trusted-ca\") pod \"console-operator-58897d9998-2pdbs\" (UID: \"07f1cf4a-506e-469a-a8e5-0adc68a5ab81\") " pod="openshift-console-operator/console-operator-58897d9998-2pdbs" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.297216 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5ba77307-01a4-43a3-8888-3da1c700ea2e-console-serving-cert\") pod \"console-f9d7485db-ctkcp\" (UID: \"5ba77307-01a4-43a3-8888-3da1c700ea2e\") " pod="openshift-console/console-f9d7485db-ctkcp" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.297398 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rtd5w"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.298630 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5j88f"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.298666 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/56722804-fa7c-4639-a4d3-5085cd965f8b-audit-policies\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.299106 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.299222 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/56722804-fa7c-4639-a4d3-5085cd965f8b-audit-dir\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.299325 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.299661 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5ba77307-01a4-43a3-8888-3da1c700ea2e-console-config\") pod \"console-f9d7485db-ctkcp\" (UID: \"5ba77307-01a4-43a3-8888-3da1c700ea2e\") " pod="openshift-console/console-f9d7485db-ctkcp" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.299970 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/da95bbc1-a1c0-428f-9884-cda886a8da3c-auth-proxy-config\") pod \"machine-approver-56656f9798-pkdzv\" (UID: \"da95bbc1-a1c0-428f-9884-cda886a8da3c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pkdzv" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.300468 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07f1cf4a-506e-469a-a8e5-0adc68a5ab81-serving-cert\") pod \"console-operator-58897d9998-2pdbs\" (UID: \"07f1cf4a-506e-469a-a8e5-0adc68a5ab81\") " pod="openshift-console-operator/console-operator-58897d9998-2pdbs" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.300596 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5ba77307-01a4-43a3-8888-3da1c700ea2e-console-oauth-config\") pod \"console-f9d7485db-ctkcp\" (UID: \"5ba77307-01a4-43a3-8888-3da1c700ea2e\") " pod="openshift-console/console-f9d7485db-ctkcp" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.300621 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.300678 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.301221 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tfc7g"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.301317 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07f1cf4a-506e-469a-a8e5-0adc68a5ab81-config\") pod \"console-operator-58897d9998-2pdbs\" (UID: \"07f1cf4a-506e-469a-a8e5-0adc68a5ab81\") " pod="openshift-console-operator/console-operator-58897d9998-2pdbs" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.301264 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.301444 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5ba77307-01a4-43a3-8888-3da1c700ea2e-trusted-ca-bundle\") pod \"console-f9d7485db-ctkcp\" (UID: \"5ba77307-01a4-43a3-8888-3da1c700ea2e\") " pod="openshift-console/console-f9d7485db-ctkcp" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.301464 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5ba77307-01a4-43a3-8888-3da1c700ea2e-oauth-serving-cert\") pod \"console-f9d7485db-ctkcp\" (UID: \"5ba77307-01a4-43a3-8888-3da1c700ea2e\") " pod="openshift-console/console-f9d7485db-ctkcp" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.302111 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/c975f343-d9c0-4145-b0b8-bdc11dbd1d66-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-mgh2f\" (UID: \"c975f343-d9c0-4145-b0b8-bdc11dbd1d66\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mgh2f" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.303030 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.303186 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-wn8jl"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.303409 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.303762 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.303950 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.304577 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.304664 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8kfz8"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.305553 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xg5l9"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.305858 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/da95bbc1-a1c0-428f-9884-cda886a8da3c-machine-approver-tls\") pod \"machine-approver-56656f9798-pkdzv\" (UID: \"da95bbc1-a1c0-428f-9884-cda886a8da3c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pkdzv" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.306599 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325780-9b22t"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.307900 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-lgfzr"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.308903 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-vtmjj"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.309947 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-nvt5z"] Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.314639 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e3b81c30-6a85-430c-a915-6322d20953ac-metrics-tls\") pod \"ingress-operator-5b745b69d9-mfp6b\" (UID: \"e3b81c30-6a85-430c-a915-6322d20953ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mfp6b" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.321874 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.347050 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.355358 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e3b81c30-6a85-430c-a915-6322d20953ac-trusted-ca\") pod \"ingress-operator-5b745b69d9-mfp6b\" (UID: \"e3b81c30-6a85-430c-a915-6322d20953ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mfp6b" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.361170 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.381030 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.401183 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.420334 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.442499 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.448971 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09503673-286e-48af-b18b-37bff87f8019-serving-cert\") pod \"authentication-operator-69f744f599-fvr9s\" (UID: \"09503673-286e-48af-b18b-37bff87f8019\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fvr9s" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.461789 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.482036 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.491052 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09503673-286e-48af-b18b-37bff87f8019-config\") pod \"authentication-operator-69f744f599-fvr9s\" (UID: \"09503673-286e-48af-b18b-37bff87f8019\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fvr9s" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.500952 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.532178 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.535512 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09503673-286e-48af-b18b-37bff87f8019-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-fvr9s\" (UID: \"09503673-286e-48af-b18b-37bff87f8019\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fvr9s" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.540739 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.561056 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.581573 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.600594 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.609420 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8725e189-8b15-4c13-ae9b-f1230476e572-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-nsspd\" (UID: \"8725e189-8b15-4c13-ae9b-f1230476e572\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nsspd" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.621548 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.631503 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8725e189-8b15-4c13-ae9b-f1230476e572-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-nsspd\" (UID: \"8725e189-8b15-4c13-ae9b-f1230476e572\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nsspd" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.641092 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.660575 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.662555 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09503673-286e-48af-b18b-37bff87f8019-service-ca-bundle\") pod \"authentication-operator-69f744f599-fvr9s\" (UID: \"09503673-286e-48af-b18b-37bff87f8019\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fvr9s" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.681113 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.701287 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.721204 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.741452 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.760721 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.801833 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.821422 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.840979 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.880795 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.900545 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.920248 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.940500 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.960532 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 04 03:12:53 crc kubenswrapper[4742]: I1004 03:12:53.981349 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.001093 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.036513 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6w26\" (UniqueName: \"kubernetes.io/projected/0062ea87-e9b5-44f9-88c2-27495348485e-kube-api-access-b6w26\") pod \"machine-api-operator-5694c8668f-fcc5d\" (UID: \"0062ea87-e9b5-44f9-88c2-27495348485e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fcc5d" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.059232 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-td4wb\" (UniqueName: \"kubernetes.io/projected/e537cd22-13ae-4f86-a246-aa53e62e8e4d-kube-api-access-td4wb\") pod \"apiserver-76f77b778f-f5978\" (UID: \"e537cd22-13ae-4f86-a246-aa53e62e8e4d\") " pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.081042 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8flcs\" (UniqueName: \"kubernetes.io/projected/690f72b6-3754-442a-b1c9-59ff8ed83c1a-kube-api-access-8flcs\") pod \"controller-manager-879f6c89f-gszdm\" (UID: \"690f72b6-3754-442a-b1c9-59ff8ed83c1a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-gszdm" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.086515 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-fcc5d" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.094960 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngndk\" (UniqueName: \"kubernetes.io/projected/938dc4e6-1064-4eab-a3e3-8e3096109655-kube-api-access-ngndk\") pod \"route-controller-manager-6576b87f9c-t7pnt\" (UID: \"938dc4e6-1064-4eab-a3e3-8e3096109655\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t7pnt" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.116844 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjkld\" (UniqueName: \"kubernetes.io/projected/f16638c6-63a8-41e3-9f0c-64fccee4ce3b-kube-api-access-fjkld\") pod \"openshift-apiserver-operator-796bbdcf4f-88xxl\" (UID: \"f16638c6-63a8-41e3-9f0c-64fccee4ce3b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-88xxl" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.121120 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.140811 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.161032 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.182065 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.201299 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.219605 4742 request.go:700] Waited for 1.012839029s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/configmaps?fieldSelector=metadata.name%3Dmarketplace-trusted-ca&limit=500&resourceVersion=0 Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.226118 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.241126 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.261337 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.281261 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.301030 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.303251 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.320534 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.328022 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-fcc5d"] Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.329701 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t7pnt" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.341107 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.361868 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.372730 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-gszdm" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.381943 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.401212 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.404102 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-88xxl" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.427919 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.441758 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.464775 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.483791 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.488154 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-f5978"] Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.501762 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.522175 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.542109 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.547190 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-t7pnt"] Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.557673 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-f5978" event={"ID":"e537cd22-13ae-4f86-a246-aa53e62e8e4d","Type":"ContainerStarted","Data":"88ddd7d6fcec8f1f9cf62de08e4b99ad5a5d7792d2d0b68a81a51e1aadd78abc"} Oct 04 03:12:54 crc kubenswrapper[4742]: W1004 03:12:54.563482 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod938dc4e6_1064_4eab_a3e3_8e3096109655.slice/crio-c103178b8682daab0e696cffcb94dc93f414347f16a5057bfe2164e47bb29ff1 WatchSource:0}: Error finding container c103178b8682daab0e696cffcb94dc93f414347f16a5057bfe2164e47bb29ff1: Status 404 returned error can't find the container with id c103178b8682daab0e696cffcb94dc93f414347f16a5057bfe2164e47bb29ff1 Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.563717 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.564532 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-fcc5d" event={"ID":"0062ea87-e9b5-44f9-88c2-27495348485e","Type":"ContainerStarted","Data":"f84fa50393d0c96f8f2d95ecfb1ec6d7fe43e627f0a49a12f968a7ace065ef87"} Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.564579 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-fcc5d" event={"ID":"0062ea87-e9b5-44f9-88c2-27495348485e","Type":"ContainerStarted","Data":"53bd0a90d1af70895d3d9e04462715efe6f9866bead906734a7a1c5848673bbf"} Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.580905 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.603467 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-gszdm"] Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.604425 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.620703 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 04 03:12:54 crc kubenswrapper[4742]: W1004 03:12:54.620752 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod690f72b6_3754_442a_b1c9_59ff8ed83c1a.slice/crio-b3c480a48a7dcbccc27a89b8e256fb821b91bfe0860bd5386db6f0644c17291d WatchSource:0}: Error finding container b3c480a48a7dcbccc27a89b8e256fb821b91bfe0860bd5386db6f0644c17291d: Status 404 returned error can't find the container with id b3c480a48a7dcbccc27a89b8e256fb821b91bfe0860bd5386db6f0644c17291d Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.637429 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-88xxl"] Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.641434 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 04 03:12:54 crc kubenswrapper[4742]: W1004 03:12:54.643876 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf16638c6_63a8_41e3_9f0c_64fccee4ce3b.slice/crio-166a47bf55f4baafe0ed583d206ee0d95199eb93efd51df0d334627786c44dab WatchSource:0}: Error finding container 166a47bf55f4baafe0ed583d206ee0d95199eb93efd51df0d334627786c44dab: Status 404 returned error can't find the container with id 166a47bf55f4baafe0ed583d206ee0d95199eb93efd51df0d334627786c44dab Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.661134 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.685706 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.701555 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.721148 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.741383 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.762066 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.781908 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.801667 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.822575 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.842152 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.860798 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.881542 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.902234 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.921848 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.943095 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.961526 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 04 03:12:54 crc kubenswrapper[4742]: I1004 03:12:54.981444 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.001658 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.020864 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.040860 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.061193 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.081485 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.100574 4742 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.121347 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.140700 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.162407 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.216836 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ls9r\" (UniqueName: \"kubernetes.io/projected/07f1cf4a-506e-469a-a8e5-0adc68a5ab81-kube-api-access-2ls9r\") pod \"console-operator-58897d9998-2pdbs\" (UID: \"07f1cf4a-506e-469a-a8e5-0adc68a5ab81\") " pod="openshift-console-operator/console-operator-58897d9998-2pdbs" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.220671 4742 request.go:700] Waited for 1.925054081s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-console/serviceaccounts/console/token Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.237436 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c975f343-d9c0-4145-b0b8-bdc11dbd1d66-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-mgh2f\" (UID: \"c975f343-d9c0-4145-b0b8-bdc11dbd1d66\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mgh2f" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.258847 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvq2q\" (UniqueName: \"kubernetes.io/projected/5ba77307-01a4-43a3-8888-3da1c700ea2e-kube-api-access-qvq2q\") pod \"console-f9d7485db-ctkcp\" (UID: \"5ba77307-01a4-43a3-8888-3da1c700ea2e\") " pod="openshift-console/console-f9d7485db-ctkcp" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.265590 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgfbg\" (UniqueName: \"kubernetes.io/projected/e3b81c30-6a85-430c-a915-6322d20953ac-kube-api-access-wgfbg\") pod \"ingress-operator-5b745b69d9-mfp6b\" (UID: \"e3b81c30-6a85-430c-a915-6322d20953ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mfp6b" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.279744 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6c8p8\" (UniqueName: \"kubernetes.io/projected/c975f343-d9c0-4145-b0b8-bdc11dbd1d66-kube-api-access-6c8p8\") pod \"cluster-image-registry-operator-dc59b4c8b-mgh2f\" (UID: \"c975f343-d9c0-4145-b0b8-bdc11dbd1d66\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mgh2f" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.300966 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9w5jf\" (UniqueName: \"kubernetes.io/projected/56722804-fa7c-4639-a4d3-5085cd965f8b-kube-api-access-9w5jf\") pod \"oauth-openshift-558db77b4-mspvb\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.316754 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e3b81c30-6a85-430c-a915-6322d20953ac-bound-sa-token\") pod \"ingress-operator-5b745b69d9-mfp6b\" (UID: \"e3b81c30-6a85-430c-a915-6322d20953ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mfp6b" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.335023 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkrcg\" (UniqueName: \"kubernetes.io/projected/0e5511f2-7587-4638-bd0c-af91c57a0163-kube-api-access-dkrcg\") pod \"downloads-7954f5f757-h22jn\" (UID: \"0e5511f2-7587-4638-bd0c-af91c57a0163\") " pod="openshift-console/downloads-7954f5f757-h22jn" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.358770 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.362397 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pd8s\" (UniqueName: \"kubernetes.io/projected/8725e189-8b15-4c13-ae9b-f1230476e572-kube-api-access-9pd8s\") pod \"kube-storage-version-migrator-operator-b67b599dd-nsspd\" (UID: \"8725e189-8b15-4c13-ae9b-f1230476e572\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nsspd" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.377234 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-2pdbs" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.381204 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqbln\" (UniqueName: \"kubernetes.io/projected/da95bbc1-a1c0-428f-9884-cda886a8da3c-kube-api-access-rqbln\") pod \"machine-approver-56656f9798-pkdzv\" (UID: \"da95bbc1-a1c0-428f-9884-cda886a8da3c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pkdzv" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.385139 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ctkcp" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.395938 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mgh2f" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.399318 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rplt\" (UniqueName: \"kubernetes.io/projected/09503673-286e-48af-b18b-37bff87f8019-kube-api-access-2rplt\") pod \"authentication-operator-69f744f599-fvr9s\" (UID: \"09503673-286e-48af-b18b-37bff87f8019\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fvr9s" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.424887 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4892c4de-41a2-4c58-bb59-36a6c584da2f-registry-tls\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.424931 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb08fe29-5abf-43af-a622-f1db8e955e55-serving-cert\") pod \"openshift-config-operator-7777fb866f-8w75v\" (UID: \"bb08fe29-5abf-43af-a622-f1db8e955e55\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8w75v" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.424979 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6d53af45-9fb2-4d36-8c21-aa167c8ed70f-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-mrq8q\" (UID: \"6d53af45-9fb2-4d36-8c21-aa167c8ed70f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.425007 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8edba022-cd27-43b2-87e3-a0e92562c450-config\") pod \"kube-controller-manager-operator-78b949d7b-rlbvb\" (UID: \"8edba022-cd27-43b2-87e3-a0e92562c450\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rlbvb" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.425031 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4892c4de-41a2-4c58-bb59-36a6c584da2f-bound-sa-token\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.425061 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7940c6dd-4863-4c92-ad0e-27abb961da4a-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-dmd8c\" (UID: \"7940c6dd-4863-4c92-ad0e-27abb961da4a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-dmd8c" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.425088 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4892c4de-41a2-4c58-bb59-36a6c584da2f-trusted-ca\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.425114 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7940c6dd-4863-4c92-ad0e-27abb961da4a-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-dmd8c\" (UID: \"7940c6dd-4863-4c92-ad0e-27abb961da4a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-dmd8c" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.425133 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/528adbb9-98ba-45d3-ba3a-2fe715e65cf6-metrics-tls\") pod \"dns-operator-744455d44c-cms9s\" (UID: \"528adbb9-98ba-45d3-ba3a-2fe715e65cf6\") " pod="openshift-dns-operator/dns-operator-744455d44c-cms9s" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.425161 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d33e79f4-4871-4e03-ac04-2000ceb3416d-etcd-service-ca\") pod \"etcd-operator-b45778765-v92dt\" (UID: \"d33e79f4-4871-4e03-ac04-2000ceb3416d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v92dt" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.425180 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6d53af45-9fb2-4d36-8c21-aa167c8ed70f-audit-dir\") pod \"apiserver-7bbb656c7d-mrq8q\" (UID: \"6d53af45-9fb2-4d36-8c21-aa167c8ed70f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.425201 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bb08fe29-5abf-43af-a622-f1db8e955e55-available-featuregates\") pod \"openshift-config-operator-7777fb866f-8w75v\" (UID: \"bb08fe29-5abf-43af-a622-f1db8e955e55\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8w75v" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.425231 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9s6j\" (UniqueName: \"kubernetes.io/projected/bb08fe29-5abf-43af-a622-f1db8e955e55-kube-api-access-h9s6j\") pod \"openshift-config-operator-7777fb866f-8w75v\" (UID: \"bb08fe29-5abf-43af-a622-f1db8e955e55\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8w75v" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.425252 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk46r\" (UniqueName: \"kubernetes.io/projected/4892c4de-41a2-4c58-bb59-36a6c584da2f-kube-api-access-tk46r\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.425298 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d33e79f4-4871-4e03-ac04-2000ceb3416d-serving-cert\") pod \"etcd-operator-b45778765-v92dt\" (UID: \"d33e79f4-4871-4e03-ac04-2000ceb3416d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v92dt" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.425320 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pft7q\" (UniqueName: \"kubernetes.io/projected/d33e79f4-4871-4e03-ac04-2000ceb3416d-kube-api-access-pft7q\") pod \"etcd-operator-b45778765-v92dt\" (UID: \"d33e79f4-4871-4e03-ac04-2000ceb3416d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v92dt" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.425462 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/62ad7c0a-a108-4c0f-b998-1accb11feb51-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-ncmp2\" (UID: \"62ad7c0a-a108-4c0f-b998-1accb11feb51\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ncmp2" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.425635 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6d53af45-9fb2-4d36-8c21-aa167c8ed70f-etcd-client\") pod \"apiserver-7bbb656c7d-mrq8q\" (UID: \"6d53af45-9fb2-4d36-8c21-aa167c8ed70f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.425707 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnxpt\" (UniqueName: \"kubernetes.io/projected/528adbb9-98ba-45d3-ba3a-2fe715e65cf6-kube-api-access-rnxpt\") pod \"dns-operator-744455d44c-cms9s\" (UID: \"528adbb9-98ba-45d3-ba3a-2fe715e65cf6\") " pod="openshift-dns-operator/dns-operator-744455d44c-cms9s" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.425823 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4892c4de-41a2-4c58-bb59-36a6c584da2f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.425891 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/6d53af45-9fb2-4d36-8c21-aa167c8ed70f-encryption-config\") pod \"apiserver-7bbb656c7d-mrq8q\" (UID: \"6d53af45-9fb2-4d36-8c21-aa167c8ed70f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.426049 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d33e79f4-4871-4e03-ac04-2000ceb3416d-etcd-ca\") pod \"etcd-operator-b45778765-v92dt\" (UID: \"d33e79f4-4871-4e03-ac04-2000ceb3416d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v92dt" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.426088 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8edba022-cd27-43b2-87e3-a0e92562c450-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-rlbvb\" (UID: \"8edba022-cd27-43b2-87e3-a0e92562c450\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rlbvb" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.426137 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b2832d6-470e-4f3d-bec4-83012fb6e541-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-6nmn2\" (UID: \"5b2832d6-470e-4f3d-bec4-83012fb6e541\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6nmn2" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.426210 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7940c6dd-4863-4c92-ad0e-27abb961da4a-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-dmd8c\" (UID: \"7940c6dd-4863-4c92-ad0e-27abb961da4a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-dmd8c" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.426376 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5b2832d6-470e-4f3d-bec4-83012fb6e541-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-6nmn2\" (UID: \"5b2832d6-470e-4f3d-bec4-83012fb6e541\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6nmn2" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.426435 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d33e79f4-4871-4e03-ac04-2000ceb3416d-etcd-client\") pod \"etcd-operator-b45778765-v92dt\" (UID: \"d33e79f4-4871-4e03-ac04-2000ceb3416d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v92dt" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.426465 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d33e79f4-4871-4e03-ac04-2000ceb3416d-config\") pod \"etcd-operator-b45778765-v92dt\" (UID: \"d33e79f4-4871-4e03-ac04-2000ceb3416d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v92dt" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.426496 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8edba022-cd27-43b2-87e3-a0e92562c450-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-rlbvb\" (UID: \"8edba022-cd27-43b2-87e3-a0e92562c450\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rlbvb" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.426521 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6d53af45-9fb2-4d36-8c21-aa167c8ed70f-audit-policies\") pod \"apiserver-7bbb656c7d-mrq8q\" (UID: \"6d53af45-9fb2-4d36-8c21-aa167c8ed70f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.426543 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/6d53af45-9fb2-4d36-8c21-aa167c8ed70f-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-mrq8q\" (UID: \"6d53af45-9fb2-4d36-8c21-aa167c8ed70f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.426579 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmrbm\" (UniqueName: \"kubernetes.io/projected/5b2832d6-470e-4f3d-bec4-83012fb6e541-kube-api-access-bmrbm\") pod \"openshift-controller-manager-operator-756b6f6bc6-6nmn2\" (UID: \"5b2832d6-470e-4f3d-bec4-83012fb6e541\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6nmn2" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.426608 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4892c4de-41a2-4c58-bb59-36a6c584da2f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.426633 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfq9w\" (UniqueName: \"kubernetes.io/projected/6d53af45-9fb2-4d36-8c21-aa167c8ed70f-kube-api-access-mfq9w\") pod \"apiserver-7bbb656c7d-mrq8q\" (UID: \"6d53af45-9fb2-4d36-8c21-aa167c8ed70f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.426667 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.426693 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2h7gh\" (UniqueName: \"kubernetes.io/projected/62ad7c0a-a108-4c0f-b998-1accb11feb51-kube-api-access-2h7gh\") pod \"control-plane-machine-set-operator-78cbb6b69f-ncmp2\" (UID: \"62ad7c0a-a108-4c0f-b998-1accb11feb51\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ncmp2" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.426900 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4892c4de-41a2-4c58-bb59-36a6c584da2f-registry-certificates\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.427181 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6d53af45-9fb2-4d36-8c21-aa167c8ed70f-serving-cert\") pod \"apiserver-7bbb656c7d-mrq8q\" (UID: \"6d53af45-9fb2-4d36-8c21-aa167c8ed70f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q" Oct 04 03:12:55 crc kubenswrapper[4742]: E1004 03:12:55.427295 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:12:55.927249499 +0000 UTC m=+141.859586685 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.430012 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-h22jn" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.494090 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mfp6b" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.504908 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-fvr9s" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.512156 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nsspd" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.531508 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.531813 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2h7gh\" (UniqueName: \"kubernetes.io/projected/62ad7c0a-a108-4c0f-b998-1accb11feb51-kube-api-access-2h7gh\") pod \"control-plane-machine-set-operator-78cbb6b69f-ncmp2\" (UID: \"62ad7c0a-a108-4c0f-b998-1accb11feb51\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ncmp2" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.531868 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4892c4de-41a2-4c58-bb59-36a6c584da2f-registry-certificates\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.531886 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6d53af45-9fb2-4d36-8c21-aa167c8ed70f-serving-cert\") pod \"apiserver-7bbb656c7d-mrq8q\" (UID: \"6d53af45-9fb2-4d36-8c21-aa167c8ed70f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.531907 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/c151b75e-6548-4d28-851e-d8f4eac68183-signing-cabundle\") pod \"service-ca-9c57cc56f-vtmjj\" (UID: \"c151b75e-6548-4d28-851e-d8f4eac68183\") " pod="openshift-service-ca/service-ca-9c57cc56f-vtmjj" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.531925 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7f11b3c3-6380-4f4d-9d03-622e66e33bc8-secret-volume\") pod \"collect-profiles-29325780-9b22t\" (UID: \"7f11b3c3-6380-4f4d-9d03-622e66e33bc8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325780-9b22t" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.531943 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4ba39c3a-d31f-4f1d-807d-c9fcf6defa34-images\") pod \"machine-config-operator-74547568cd-6prnz\" (UID: \"4ba39c3a-d31f-4f1d-807d-c9fcf6defa34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6prnz" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.531961 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/093665e3-6dc0-4c51-a0f3-71363eef8367-service-ca-bundle\") pod \"router-default-5444994796-85fhx\" (UID: \"093665e3-6dc0-4c51-a0f3-71363eef8367\") " pod="openshift-ingress/router-default-5444994796-85fhx" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.531976 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3a41af56-8a6d-49ba-8b85-1c546d49c198-config-volume\") pod \"dns-default-nvt5z\" (UID: \"3a41af56-8a6d-49ba-8b85-1c546d49c198\") " pod="openshift-dns/dns-default-nvt5z" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.532012 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb08fe29-5abf-43af-a622-f1db8e955e55-serving-cert\") pod \"openshift-config-operator-7777fb866f-8w75v\" (UID: \"bb08fe29-5abf-43af-a622-f1db8e955e55\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8w75v" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.532028 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8edba022-cd27-43b2-87e3-a0e92562c450-config\") pod \"kube-controller-manager-operator-78b949d7b-rlbvb\" (UID: \"8edba022-cd27-43b2-87e3-a0e92562c450\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rlbvb" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.532044 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6d53af45-9fb2-4d36-8c21-aa167c8ed70f-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-mrq8q\" (UID: \"6d53af45-9fb2-4d36-8c21-aa167c8ed70f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.532072 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/77153235-6efa-4db2-a4a3-5153d9ba3568-node-bootstrap-token\") pod \"machine-config-server-4v795\" (UID: \"77153235-6efa-4db2-a4a3-5153d9ba3568\") " pod="openshift-machine-config-operator/machine-config-server-4v795" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.532099 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/093665e3-6dc0-4c51-a0f3-71363eef8367-default-certificate\") pod \"router-default-5444994796-85fhx\" (UID: \"093665e3-6dc0-4c51-a0f3-71363eef8367\") " pod="openshift-ingress/router-default-5444994796-85fhx" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.532124 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/528adbb9-98ba-45d3-ba3a-2fe715e65cf6-metrics-tls\") pod \"dns-operator-744455d44c-cms9s\" (UID: \"528adbb9-98ba-45d3-ba3a-2fe715e65cf6\") " pod="openshift-dns-operator/dns-operator-744455d44c-cms9s" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.532152 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/35542dc5-96db-4eca-a2c2-85ebf65ea31f-socket-dir\") pod \"csi-hostpathplugin-wn8jl\" (UID: \"35542dc5-96db-4eca-a2c2-85ebf65ea31f\") " pod="hostpath-provisioner/csi-hostpathplugin-wn8jl" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.532172 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d33e79f4-4871-4e03-ac04-2000ceb3416d-etcd-service-ca\") pod \"etcd-operator-b45778765-v92dt\" (UID: \"d33e79f4-4871-4e03-ac04-2000ceb3416d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v92dt" Oct 04 03:12:55 crc kubenswrapper[4742]: E1004 03:12:55.532564 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:12:56.032540701 +0000 UTC m=+141.964877877 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.533165 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8edba022-cd27-43b2-87e3-a0e92562c450-config\") pod \"kube-controller-manager-operator-78b949d7b-rlbvb\" (UID: \"8edba022-cd27-43b2-87e3-a0e92562c450\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rlbvb" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.533477 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6d53af45-9fb2-4d36-8c21-aa167c8ed70f-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-mrq8q\" (UID: \"6d53af45-9fb2-4d36-8c21-aa167c8ed70f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.532219 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bb08fe29-5abf-43af-a622-f1db8e955e55-available-featuregates\") pod \"openshift-config-operator-7777fb866f-8w75v\" (UID: \"bb08fe29-5abf-43af-a622-f1db8e955e55\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8w75v" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.533867 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a8b41c08-4e89-4f08-8d45-256c306fbea0-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tfc7g\" (UID: \"a8b41c08-4e89-4f08-8d45-256c306fbea0\") " pod="openshift-marketplace/marketplace-operator-79b997595-tfc7g" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.533920 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3a41af56-8a6d-49ba-8b85-1c546d49c198-metrics-tls\") pod \"dns-default-nvt5z\" (UID: \"3a41af56-8a6d-49ba-8b85-1c546d49c198\") " pod="openshift-dns/dns-default-nvt5z" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.533994 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d33e79f4-4871-4e03-ac04-2000ceb3416d-serving-cert\") pod \"etcd-operator-b45778765-v92dt\" (UID: \"d33e79f4-4871-4e03-ac04-2000ceb3416d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v92dt" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.534060 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d33e79f4-4871-4e03-ac04-2000ceb3416d-etcd-service-ca\") pod \"etcd-operator-b45778765-v92dt\" (UID: \"d33e79f4-4871-4e03-ac04-2000ceb3416d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v92dt" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.534170 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pft7q\" (UniqueName: \"kubernetes.io/projected/d33e79f4-4871-4e03-ac04-2000ceb3416d-kube-api-access-pft7q\") pod \"etcd-operator-b45778765-v92dt\" (UID: \"d33e79f4-4871-4e03-ac04-2000ceb3416d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v92dt" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.534305 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bb08fe29-5abf-43af-a622-f1db8e955e55-available-featuregates\") pod \"openshift-config-operator-7777fb866f-8w75v\" (UID: \"bb08fe29-5abf-43af-a622-f1db8e955e55\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8w75v" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.534389 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6d53af45-9fb2-4d36-8c21-aa167c8ed70f-etcd-client\") pod \"apiserver-7bbb656c7d-mrq8q\" (UID: \"6d53af45-9fb2-4d36-8c21-aa167c8ed70f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.534421 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/912dc882-2c54-4a47-b26b-6f3a63f9f33f-webhook-cert\") pod \"packageserver-d55dfcdfc-rtd5w\" (UID: \"912dc882-2c54-4a47-b26b-6f3a63f9f33f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rtd5w" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.534519 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4892c4de-41a2-4c58-bb59-36a6c584da2f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.534546 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/6d53af45-9fb2-4d36-8c21-aa167c8ed70f-encryption-config\") pod \"apiserver-7bbb656c7d-mrq8q\" (UID: \"6d53af45-9fb2-4d36-8c21-aa167c8ed70f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.534568 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blwh2\" (UniqueName: \"kubernetes.io/projected/530fb6b6-9c36-4d13-8aa9-903d7080082f-kube-api-access-blwh2\") pod \"machine-config-controller-84d6567774-krpzs\" (UID: \"530fb6b6-9c36-4d13-8aa9-903d7080082f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-krpzs" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.534589 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b16d0ad2-0b45-45fe-bc2c-215697447010-config\") pod \"service-ca-operator-777779d784-ktnrj\" (UID: \"b16d0ad2-0b45-45fe-bc2c-215697447010\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ktnrj" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.534761 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/35542dc5-96db-4eca-a2c2-85ebf65ea31f-csi-data-dir\") pod \"csi-hostpathplugin-wn8jl\" (UID: \"35542dc5-96db-4eca-a2c2-85ebf65ea31f\") " pod="hostpath-provisioner/csi-hostpathplugin-wn8jl" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.534800 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d33e79f4-4871-4e03-ac04-2000ceb3416d-etcd-ca\") pod \"etcd-operator-b45778765-v92dt\" (UID: \"d33e79f4-4871-4e03-ac04-2000ceb3416d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v92dt" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.534873 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8edba022-cd27-43b2-87e3-a0e92562c450-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-rlbvb\" (UID: \"8edba022-cd27-43b2-87e3-a0e92562c450\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rlbvb" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.534923 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5b2832d6-470e-4f3d-bec4-83012fb6e541-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-6nmn2\" (UID: \"5b2832d6-470e-4f3d-bec4-83012fb6e541\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6nmn2" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.534951 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/093665e3-6dc0-4c51-a0f3-71363eef8367-metrics-certs\") pod \"router-default-5444994796-85fhx\" (UID: \"093665e3-6dc0-4c51-a0f3-71363eef8367\") " pod="openshift-ingress/router-default-5444994796-85fhx" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.534986 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/c151b75e-6548-4d28-851e-d8f4eac68183-signing-key\") pod \"service-ca-9c57cc56f-vtmjj\" (UID: \"c151b75e-6548-4d28-851e-d8f4eac68183\") " pod="openshift-service-ca/service-ca-9c57cc56f-vtmjj" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.535090 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8edba022-cd27-43b2-87e3-a0e92562c450-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-rlbvb\" (UID: \"8edba022-cd27-43b2-87e3-a0e92562c450\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rlbvb" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.535111 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpsq6\" (UniqueName: \"kubernetes.io/projected/9d4974fb-9d3c-4e2d-ac54-7c3d4b4b3f89-kube-api-access-wpsq6\") pod \"cluster-samples-operator-665b6dd947-qznqm\" (UID: \"9d4974fb-9d3c-4e2d-ac54-7c3d4b4b3f89\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qznqm" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.535192 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/6d53af45-9fb2-4d36-8c21-aa167c8ed70f-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-mrq8q\" (UID: \"6d53af45-9fb2-4d36-8c21-aa167c8ed70f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.535212 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/adc0635a-397c-48ba-b103-a8d6e62ae535-srv-cert\") pod \"olm-operator-6b444d44fb-5j88f\" (UID: \"adc0635a-397c-48ba-b103-a8d6e62ae535\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5j88f" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.535264 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmrbm\" (UniqueName: \"kubernetes.io/projected/5b2832d6-470e-4f3d-bec4-83012fb6e541-kube-api-access-bmrbm\") pod \"openshift-controller-manager-operator-756b6f6bc6-6nmn2\" (UID: \"5b2832d6-470e-4f3d-bec4-83012fb6e541\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6nmn2" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.535407 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvwhg\" (UniqueName: \"kubernetes.io/projected/7336d2fb-f015-4e22-8419-fdefc49f6137-kube-api-access-dvwhg\") pod \"catalog-operator-68c6474976-8kfz8\" (UID: \"7336d2fb-f015-4e22-8419-fdefc49f6137\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8kfz8" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.535427 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88cbx\" (UniqueName: \"kubernetes.io/projected/3a41af56-8a6d-49ba-8b85-1c546d49c198-kube-api-access-88cbx\") pod \"dns-default-nvt5z\" (UID: \"3a41af56-8a6d-49ba-8b85-1c546d49c198\") " pod="openshift-dns/dns-default-nvt5z" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.535663 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4892c4de-41a2-4c58-bb59-36a6c584da2f-registry-certificates\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.535925 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d33e79f4-4871-4e03-ac04-2000ceb3416d-etcd-ca\") pod \"etcd-operator-b45778765-v92dt\" (UID: \"d33e79f4-4871-4e03-ac04-2000ceb3416d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v92dt" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.536471 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4892c4de-41a2-4c58-bb59-36a6c584da2f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.536500 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfq9w\" (UniqueName: \"kubernetes.io/projected/6d53af45-9fb2-4d36-8c21-aa167c8ed70f-kube-api-access-mfq9w\") pod \"apiserver-7bbb656c7d-mrq8q\" (UID: \"6d53af45-9fb2-4d36-8c21-aa167c8ed70f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.536525 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x55xr\" (UniqueName: \"kubernetes.io/projected/a8b41c08-4e89-4f08-8d45-256c306fbea0-kube-api-access-x55xr\") pod \"marketplace-operator-79b997595-tfc7g\" (UID: \"a8b41c08-4e89-4f08-8d45-256c306fbea0\") " pod="openshift-marketplace/marketplace-operator-79b997595-tfc7g" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.536863 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4892c4de-41a2-4c58-bb59-36a6c584da2f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.536924 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/6d53af45-9fb2-4d36-8c21-aa167c8ed70f-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-mrq8q\" (UID: \"6d53af45-9fb2-4d36-8c21-aa167c8ed70f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.536934 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a8b41c08-4e89-4f08-8d45-256c306fbea0-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tfc7g\" (UID: \"a8b41c08-4e89-4f08-8d45-256c306fbea0\") " pod="openshift-marketplace/marketplace-operator-79b997595-tfc7g" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.537014 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b16d0ad2-0b45-45fe-bc2c-215697447010-serving-cert\") pod \"service-ca-operator-777779d784-ktnrj\" (UID: \"b16d0ad2-0b45-45fe-bc2c-215697447010\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ktnrj" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.537228 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/80a23793-4622-4288-a5c0-de5199e1e767-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-xg5l9\" (UID: \"80a23793-4622-4288-a5c0-de5199e1e767\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xg5l9" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.537259 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/912dc882-2c54-4a47-b26b-6f3a63f9f33f-tmpfs\") pod \"packageserver-d55dfcdfc-rtd5w\" (UID: \"912dc882-2c54-4a47-b26b-6f3a63f9f33f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rtd5w" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.537342 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fw6b\" (UniqueName: \"kubernetes.io/projected/bf0e3a33-4470-430f-8ebc-71d52b649684-kube-api-access-4fw6b\") pod \"migrator-59844c95c7-p559m\" (UID: \"bf0e3a33-4470-430f-8ebc-71d52b649684\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-p559m" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.537364 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stj62\" (UniqueName: \"kubernetes.io/projected/80a23793-4622-4288-a5c0-de5199e1e767-kube-api-access-stj62\") pod \"package-server-manager-789f6589d5-xg5l9\" (UID: \"80a23793-4622-4288-a5c0-de5199e1e767\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xg5l9" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.537432 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/093665e3-6dc0-4c51-a0f3-71363eef8367-stats-auth\") pod \"router-default-5444994796-85fhx\" (UID: \"093665e3-6dc0-4c51-a0f3-71363eef8367\") " pod="openshift-ingress/router-default-5444994796-85fhx" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.537597 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4892c4de-41a2-4c58-bb59-36a6c584da2f-registry-tls\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.537621 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7336d2fb-f015-4e22-8419-fdefc49f6137-srv-cert\") pod \"catalog-operator-68c6474976-8kfz8\" (UID: \"7336d2fb-f015-4e22-8419-fdefc49f6137\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8kfz8" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.537641 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzfch\" (UniqueName: \"kubernetes.io/projected/b16d0ad2-0b45-45fe-bc2c-215697447010-kube-api-access-mzfch\") pod \"service-ca-operator-777779d784-ktnrj\" (UID: \"b16d0ad2-0b45-45fe-bc2c-215697447010\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ktnrj" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.537737 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4892c4de-41a2-4c58-bb59-36a6c584da2f-bound-sa-token\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.537758 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7940c6dd-4863-4c92-ad0e-27abb961da4a-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-dmd8c\" (UID: \"7940c6dd-4863-4c92-ad0e-27abb961da4a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-dmd8c" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.537795 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4892c4de-41a2-4c58-bb59-36a6c584da2f-trusted-ca\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.540449 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5057f820-3d60-43e3-9cb8-ceac9650621d-config\") pod \"kube-apiserver-operator-766d6c64bb-kfp77\" (UID: \"5057f820-3d60-43e3-9cb8-ceac9650621d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kfp77" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.540495 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/77153235-6efa-4db2-a4a3-5153d9ba3568-certs\") pod \"machine-config-server-4v795\" (UID: \"77153235-6efa-4db2-a4a3-5153d9ba3568\") " pod="openshift-machine-config-operator/machine-config-server-4v795" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.540564 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7940c6dd-4863-4c92-ad0e-27abb961da4a-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-dmd8c\" (UID: \"7940c6dd-4863-4c92-ad0e-27abb961da4a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-dmd8c" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.540634 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/cd220c67-e881-4180-b1c4-4064be0794ca-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-k77jn\" (UID: \"cd220c67-e881-4180-b1c4-4064be0794ca\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-k77jn" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.540662 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/adc0635a-397c-48ba-b103-a8d6e62ae535-profile-collector-cert\") pod \"olm-operator-6b444d44fb-5j88f\" (UID: \"adc0635a-397c-48ba-b103-a8d6e62ae535\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5j88f" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.540685 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/912dc882-2c54-4a47-b26b-6f3a63f9f33f-apiservice-cert\") pod \"packageserver-d55dfcdfc-rtd5w\" (UID: \"912dc882-2c54-4a47-b26b-6f3a63f9f33f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rtd5w" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.540711 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66ddt\" (UniqueName: \"kubernetes.io/projected/7f11b3c3-6380-4f4d-9d03-622e66e33bc8-kube-api-access-66ddt\") pod \"collect-profiles-29325780-9b22t\" (UID: \"7f11b3c3-6380-4f4d-9d03-622e66e33bc8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325780-9b22t" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.540761 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6d53af45-9fb2-4d36-8c21-aa167c8ed70f-audit-dir\") pod \"apiserver-7bbb656c7d-mrq8q\" (UID: \"6d53af45-9fb2-4d36-8c21-aa167c8ed70f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.540783 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4ba39c3a-d31f-4f1d-807d-c9fcf6defa34-proxy-tls\") pod \"machine-config-operator-74547568cd-6prnz\" (UID: \"4ba39c3a-d31f-4f1d-807d-c9fcf6defa34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6prnz" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.540813 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9s6j\" (UniqueName: \"kubernetes.io/projected/bb08fe29-5abf-43af-a622-f1db8e955e55-kube-api-access-h9s6j\") pod \"openshift-config-operator-7777fb866f-8w75v\" (UID: \"bb08fe29-5abf-43af-a622-f1db8e955e55\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8w75v" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.540838 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4ba39c3a-d31f-4f1d-807d-c9fcf6defa34-auth-proxy-config\") pod \"machine-config-operator-74547568cd-6prnz\" (UID: \"4ba39c3a-d31f-4f1d-807d-c9fcf6defa34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6prnz" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.540863 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkntl\" (UniqueName: \"kubernetes.io/projected/c151b75e-6548-4d28-851e-d8f4eac68183-kube-api-access-bkntl\") pod \"service-ca-9c57cc56f-vtmjj\" (UID: \"c151b75e-6548-4d28-851e-d8f4eac68183\") " pod="openshift-service-ca/service-ca-9c57cc56f-vtmjj" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.543002 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4892c4de-41a2-4c58-bb59-36a6c584da2f-trusted-ca\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.543327 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6d53af45-9fb2-4d36-8c21-aa167c8ed70f-audit-dir\") pod \"apiserver-7bbb656c7d-mrq8q\" (UID: \"6d53af45-9fb2-4d36-8c21-aa167c8ed70f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.543632 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk46r\" (UniqueName: \"kubernetes.io/projected/4892c4de-41a2-4c58-bb59-36a6c584da2f-kube-api-access-tk46r\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.543665 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/62ad7c0a-a108-4c0f-b998-1accb11feb51-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-ncmp2\" (UID: \"62ad7c0a-a108-4c0f-b998-1accb11feb51\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ncmp2" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.543908 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/9d4974fb-9d3c-4e2d-ac54-7c3d4b4b3f89-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-qznqm\" (UID: \"9d4974fb-9d3c-4e2d-ac54-7c3d4b4b3f89\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qznqm" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.544045 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9x4m2\" (UniqueName: \"kubernetes.io/projected/35542dc5-96db-4eca-a2c2-85ebf65ea31f-kube-api-access-9x4m2\") pod \"csi-hostpathplugin-wn8jl\" (UID: \"35542dc5-96db-4eca-a2c2-85ebf65ea31f\") " pod="hostpath-provisioner/csi-hostpathplugin-wn8jl" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.546128 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnxpt\" (UniqueName: \"kubernetes.io/projected/528adbb9-98ba-45d3-ba3a-2fe715e65cf6-kube-api-access-rnxpt\") pod \"dns-operator-744455d44c-cms9s\" (UID: \"528adbb9-98ba-45d3-ba3a-2fe715e65cf6\") " pod="openshift-dns-operator/dns-operator-744455d44c-cms9s" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.546255 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/530fb6b6-9c36-4d13-8aa9-903d7080082f-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-krpzs\" (UID: \"530fb6b6-9c36-4d13-8aa9-903d7080082f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-krpzs" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.546318 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5057f820-3d60-43e3-9cb8-ceac9650621d-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-kfp77\" (UID: \"5057f820-3d60-43e3-9cb8-ceac9650621d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kfp77" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.546344 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbhrp\" (UniqueName: \"kubernetes.io/projected/4ba39c3a-d31f-4f1d-807d-c9fcf6defa34-kube-api-access-xbhrp\") pod \"machine-config-operator-74547568cd-6prnz\" (UID: \"4ba39c3a-d31f-4f1d-807d-c9fcf6defa34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6prnz" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.546420 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vztvs\" (UniqueName: \"kubernetes.io/projected/cd220c67-e881-4180-b1c4-4064be0794ca-kube-api-access-vztvs\") pod \"multus-admission-controller-857f4d67dd-k77jn\" (UID: \"cd220c67-e881-4180-b1c4-4064be0794ca\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-k77jn" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.546478 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b2832d6-470e-4f3d-bec4-83012fb6e541-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-6nmn2\" (UID: \"5b2832d6-470e-4f3d-bec4-83012fb6e541\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6nmn2" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.546507 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/35542dc5-96db-4eca-a2c2-85ebf65ea31f-registration-dir\") pod \"csi-hostpathplugin-wn8jl\" (UID: \"35542dc5-96db-4eca-a2c2-85ebf65ea31f\") " pod="hostpath-provisioner/csi-hostpathplugin-wn8jl" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.546532 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xl64r\" (UniqueName: \"kubernetes.io/projected/093665e3-6dc0-4c51-a0f3-71363eef8367-kube-api-access-xl64r\") pod \"router-default-5444994796-85fhx\" (UID: \"093665e3-6dc0-4c51-a0f3-71363eef8367\") " pod="openshift-ingress/router-default-5444994796-85fhx" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.546612 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2556513e-2080-4ed1-99cc-bb1e15a75a8e-cert\") pod \"ingress-canary-lgfzr\" (UID: \"2556513e-2080-4ed1-99cc-bb1e15a75a8e\") " pod="openshift-ingress-canary/ingress-canary-lgfzr" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.546637 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7336d2fb-f015-4e22-8419-fdefc49f6137-profile-collector-cert\") pod \"catalog-operator-68c6474976-8kfz8\" (UID: \"7336d2fb-f015-4e22-8419-fdefc49f6137\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8kfz8" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.546688 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7940c6dd-4863-4c92-ad0e-27abb961da4a-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-dmd8c\" (UID: \"7940c6dd-4863-4c92-ad0e-27abb961da4a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-dmd8c" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.546719 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnlxv\" (UniqueName: \"kubernetes.io/projected/adc0635a-397c-48ba-b103-a8d6e62ae535-kube-api-access-jnlxv\") pod \"olm-operator-6b444d44fb-5j88f\" (UID: \"adc0635a-397c-48ba-b103-a8d6e62ae535\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5j88f" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.547339 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d33e79f4-4871-4e03-ac04-2000ceb3416d-etcd-client\") pod \"etcd-operator-b45778765-v92dt\" (UID: \"d33e79f4-4871-4e03-ac04-2000ceb3416d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v92dt" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.547396 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r679r\" (UniqueName: \"kubernetes.io/projected/2556513e-2080-4ed1-99cc-bb1e15a75a8e-kube-api-access-r679r\") pod \"ingress-canary-lgfzr\" (UID: \"2556513e-2080-4ed1-99cc-bb1e15a75a8e\") " pod="openshift-ingress-canary/ingress-canary-lgfzr" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.547557 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d33e79f4-4871-4e03-ac04-2000ceb3416d-config\") pod \"etcd-operator-b45778765-v92dt\" (UID: \"d33e79f4-4871-4e03-ac04-2000ceb3416d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v92dt" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.548106 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d33e79f4-4871-4e03-ac04-2000ceb3416d-config\") pod \"etcd-operator-b45778765-v92dt\" (UID: \"d33e79f4-4871-4e03-ac04-2000ceb3416d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v92dt" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.548318 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/530fb6b6-9c36-4d13-8aa9-903d7080082f-proxy-tls\") pod \"machine-config-controller-84d6567774-krpzs\" (UID: \"530fb6b6-9c36-4d13-8aa9-903d7080082f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-krpzs" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.548416 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/35542dc5-96db-4eca-a2c2-85ebf65ea31f-plugins-dir\") pod \"csi-hostpathplugin-wn8jl\" (UID: \"35542dc5-96db-4eca-a2c2-85ebf65ea31f\") " pod="hostpath-provisioner/csi-hostpathplugin-wn8jl" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.548471 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7f11b3c3-6380-4f4d-9d03-622e66e33bc8-config-volume\") pod \"collect-profiles-29325780-9b22t\" (UID: \"7f11b3c3-6380-4f4d-9d03-622e66e33bc8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325780-9b22t" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.548649 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6d53af45-9fb2-4d36-8c21-aa167c8ed70f-audit-policies\") pod \"apiserver-7bbb656c7d-mrq8q\" (UID: \"6d53af45-9fb2-4d36-8c21-aa167c8ed70f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.548755 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/35542dc5-96db-4eca-a2c2-85ebf65ea31f-mountpoint-dir\") pod \"csi-hostpathplugin-wn8jl\" (UID: \"35542dc5-96db-4eca-a2c2-85ebf65ea31f\") " pod="hostpath-provisioner/csi-hostpathplugin-wn8jl" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.548831 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v42w5\" (UniqueName: \"kubernetes.io/projected/77153235-6efa-4db2-a4a3-5153d9ba3568-kube-api-access-v42w5\") pod \"machine-config-server-4v795\" (UID: \"77153235-6efa-4db2-a4a3-5153d9ba3568\") " pod="openshift-machine-config-operator/machine-config-server-4v795" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.548826 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7940c6dd-4863-4c92-ad0e-27abb961da4a-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-dmd8c\" (UID: \"7940c6dd-4863-4c92-ad0e-27abb961da4a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-dmd8c" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.548871 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ks892\" (UniqueName: \"kubernetes.io/projected/912dc882-2c54-4a47-b26b-6f3a63f9f33f-kube-api-access-ks892\") pod \"packageserver-d55dfcdfc-rtd5w\" (UID: \"912dc882-2c54-4a47-b26b-6f3a63f9f33f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rtd5w" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.548964 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5057f820-3d60-43e3-9cb8-ceac9650621d-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-kfp77\" (UID: \"5057f820-3d60-43e3-9cb8-ceac9650621d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kfp77" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.549201 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6d53af45-9fb2-4d36-8c21-aa167c8ed70f-audit-policies\") pod \"apiserver-7bbb656c7d-mrq8q\" (UID: \"6d53af45-9fb2-4d36-8c21-aa167c8ed70f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.551144 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b2832d6-470e-4f3d-bec4-83012fb6e541-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-6nmn2\" (UID: \"5b2832d6-470e-4f3d-bec4-83012fb6e541\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6nmn2" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.551882 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/6d53af45-9fb2-4d36-8c21-aa167c8ed70f-encryption-config\") pod \"apiserver-7bbb656c7d-mrq8q\" (UID: \"6d53af45-9fb2-4d36-8c21-aa167c8ed70f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.561359 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5b2832d6-470e-4f3d-bec4-83012fb6e541-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-6nmn2\" (UID: \"5b2832d6-470e-4f3d-bec4-83012fb6e541\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6nmn2" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.561440 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/528adbb9-98ba-45d3-ba3a-2fe715e65cf6-metrics-tls\") pod \"dns-operator-744455d44c-cms9s\" (UID: \"528adbb9-98ba-45d3-ba3a-2fe715e65cf6\") " pod="openshift-dns-operator/dns-operator-744455d44c-cms9s" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.561425 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d33e79f4-4871-4e03-ac04-2000ceb3416d-serving-cert\") pod \"etcd-operator-b45778765-v92dt\" (UID: \"d33e79f4-4871-4e03-ac04-2000ceb3416d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v92dt" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.561732 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8edba022-cd27-43b2-87e3-a0e92562c450-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-rlbvb\" (UID: \"8edba022-cd27-43b2-87e3-a0e92562c450\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rlbvb" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.561783 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6d53af45-9fb2-4d36-8c21-aa167c8ed70f-etcd-client\") pod \"apiserver-7bbb656c7d-mrq8q\" (UID: \"6d53af45-9fb2-4d36-8c21-aa167c8ed70f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.561822 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6d53af45-9fb2-4d36-8c21-aa167c8ed70f-serving-cert\") pod \"apiserver-7bbb656c7d-mrq8q\" (UID: \"6d53af45-9fb2-4d36-8c21-aa167c8ed70f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.562000 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4892c4de-41a2-4c58-bb59-36a6c584da2f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.562043 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d33e79f4-4871-4e03-ac04-2000ceb3416d-etcd-client\") pod \"etcd-operator-b45778765-v92dt\" (UID: \"d33e79f4-4871-4e03-ac04-2000ceb3416d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v92dt" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.562249 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4892c4de-41a2-4c58-bb59-36a6c584da2f-registry-tls\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.562477 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7940c6dd-4863-4c92-ad0e-27abb961da4a-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-dmd8c\" (UID: \"7940c6dd-4863-4c92-ad0e-27abb961da4a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-dmd8c" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.562924 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb08fe29-5abf-43af-a622-f1db8e955e55-serving-cert\") pod \"openshift-config-operator-7777fb866f-8w75v\" (UID: \"bb08fe29-5abf-43af-a622-f1db8e955e55\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8w75v" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.563905 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/62ad7c0a-a108-4c0f-b998-1accb11feb51-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-ncmp2\" (UID: \"62ad7c0a-a108-4c0f-b998-1accb11feb51\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ncmp2" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.565259 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2h7gh\" (UniqueName: \"kubernetes.io/projected/62ad7c0a-a108-4c0f-b998-1accb11feb51-kube-api-access-2h7gh\") pod \"control-plane-machine-set-operator-78cbb6b69f-ncmp2\" (UID: \"62ad7c0a-a108-4c0f-b998-1accb11feb51\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ncmp2" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.572242 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-fcc5d" event={"ID":"0062ea87-e9b5-44f9-88c2-27495348485e","Type":"ContainerStarted","Data":"e10ddff19d977cce1c55d4039a30e4534701f3df4298410b10cee7f67423a422"} Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.573534 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t7pnt" event={"ID":"938dc4e6-1064-4eab-a3e3-8e3096109655","Type":"ContainerStarted","Data":"61b050e50e078ef054514389b9665c66a71facfe81ec7126e3dde6a61bd818e7"} Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.573563 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t7pnt" event={"ID":"938dc4e6-1064-4eab-a3e3-8e3096109655","Type":"ContainerStarted","Data":"c103178b8682daab0e696cffcb94dc93f414347f16a5057bfe2164e47bb29ff1"} Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.574224 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t7pnt" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.575591 4742 generic.go:334] "Generic (PLEG): container finished" podID="e537cd22-13ae-4f86-a246-aa53e62e8e4d" containerID="fd1871f5c5cfb14d4357d010a0db25fce8f0056c9e442ed18fe186aa03908376" exitCode=0 Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.575641 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-f5978" event={"ID":"e537cd22-13ae-4f86-a246-aa53e62e8e4d","Type":"ContainerDied","Data":"fd1871f5c5cfb14d4357d010a0db25fce8f0056c9e442ed18fe186aa03908376"} Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.590838 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-88xxl" event={"ID":"f16638c6-63a8-41e3-9f0c-64fccee4ce3b","Type":"ContainerStarted","Data":"4ed468f8d17c6723c9f238bb2a4ddda83a97e963777b08b420ea06b2c243302b"} Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.590886 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-88xxl" event={"ID":"f16638c6-63a8-41e3-9f0c-64fccee4ce3b","Type":"ContainerStarted","Data":"166a47bf55f4baafe0ed583d206ee0d95199eb93efd51df0d334627786c44dab"} Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.592542 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-gszdm" event={"ID":"690f72b6-3754-442a-b1c9-59ff8ed83c1a","Type":"ContainerStarted","Data":"d0849799409f769f9737527b6af9087a31e3db241ce43e7c56a8db8797e9a877"} Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.592564 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-gszdm" event={"ID":"690f72b6-3754-442a-b1c9-59ff8ed83c1a","Type":"ContainerStarted","Data":"b3c480a48a7dcbccc27a89b8e256fb821b91bfe0860bd5386db6f0644c17291d"} Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.593330 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-gszdm" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.604095 4742 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-gszdm container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.604135 4742 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-gszdm" podUID="690f72b6-3754-442a-b1c9-59ff8ed83c1a" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.613752 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pft7q\" (UniqueName: \"kubernetes.io/projected/d33e79f4-4871-4e03-ac04-2000ceb3416d-kube-api-access-pft7q\") pod \"etcd-operator-b45778765-v92dt\" (UID: \"d33e79f4-4871-4e03-ac04-2000ceb3416d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v92dt" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.626241 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmrbm\" (UniqueName: \"kubernetes.io/projected/5b2832d6-470e-4f3d-bec4-83012fb6e541-kube-api-access-bmrbm\") pod \"openshift-controller-manager-operator-756b6f6bc6-6nmn2\" (UID: \"5b2832d6-470e-4f3d-bec4-83012fb6e541\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6nmn2" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.641166 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfq9w\" (UniqueName: \"kubernetes.io/projected/6d53af45-9fb2-4d36-8c21-aa167c8ed70f-kube-api-access-mfq9w\") pod \"apiserver-7bbb656c7d-mrq8q\" (UID: \"6d53af45-9fb2-4d36-8c21-aa167c8ed70f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.653285 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/093665e3-6dc0-4c51-a0f3-71363eef8367-default-certificate\") pod \"router-default-5444994796-85fhx\" (UID: \"093665e3-6dc0-4c51-a0f3-71363eef8367\") " pod="openshift-ingress/router-default-5444994796-85fhx" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.653331 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/35542dc5-96db-4eca-a2c2-85ebf65ea31f-socket-dir\") pod \"csi-hostpathplugin-wn8jl\" (UID: \"35542dc5-96db-4eca-a2c2-85ebf65ea31f\") " pod="hostpath-provisioner/csi-hostpathplugin-wn8jl" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.653353 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a8b41c08-4e89-4f08-8d45-256c306fbea0-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tfc7g\" (UID: \"a8b41c08-4e89-4f08-8d45-256c306fbea0\") " pod="openshift-marketplace/marketplace-operator-79b997595-tfc7g" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.653371 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3a41af56-8a6d-49ba-8b85-1c546d49c198-metrics-tls\") pod \"dns-default-nvt5z\" (UID: \"3a41af56-8a6d-49ba-8b85-1c546d49c198\") " pod="openshift-dns/dns-default-nvt5z" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.653408 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/912dc882-2c54-4a47-b26b-6f3a63f9f33f-webhook-cert\") pod \"packageserver-d55dfcdfc-rtd5w\" (UID: \"912dc882-2c54-4a47-b26b-6f3a63f9f33f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rtd5w" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.653426 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blwh2\" (UniqueName: \"kubernetes.io/projected/530fb6b6-9c36-4d13-8aa9-903d7080082f-kube-api-access-blwh2\") pod \"machine-config-controller-84d6567774-krpzs\" (UID: \"530fb6b6-9c36-4d13-8aa9-903d7080082f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-krpzs" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.653444 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b16d0ad2-0b45-45fe-bc2c-215697447010-config\") pod \"service-ca-operator-777779d784-ktnrj\" (UID: \"b16d0ad2-0b45-45fe-bc2c-215697447010\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ktnrj" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.653459 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/35542dc5-96db-4eca-a2c2-85ebf65ea31f-csi-data-dir\") pod \"csi-hostpathplugin-wn8jl\" (UID: \"35542dc5-96db-4eca-a2c2-85ebf65ea31f\") " pod="hostpath-provisioner/csi-hostpathplugin-wn8jl" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.654212 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/35542dc5-96db-4eca-a2c2-85ebf65ea31f-csi-data-dir\") pod \"csi-hostpathplugin-wn8jl\" (UID: \"35542dc5-96db-4eca-a2c2-85ebf65ea31f\") " pod="hostpath-provisioner/csi-hostpathplugin-wn8jl" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.654538 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/093665e3-6dc0-4c51-a0f3-71363eef8367-metrics-certs\") pod \"router-default-5444994796-85fhx\" (UID: \"093665e3-6dc0-4c51-a0f3-71363eef8367\") " pod="openshift-ingress/router-default-5444994796-85fhx" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.654570 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/c151b75e-6548-4d28-851e-d8f4eac68183-signing-key\") pod \"service-ca-9c57cc56f-vtmjj\" (UID: \"c151b75e-6548-4d28-851e-d8f4eac68183\") " pod="openshift-service-ca/service-ca-9c57cc56f-vtmjj" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.654597 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpsq6\" (UniqueName: \"kubernetes.io/projected/9d4974fb-9d3c-4e2d-ac54-7c3d4b4b3f89-kube-api-access-wpsq6\") pod \"cluster-samples-operator-665b6dd947-qznqm\" (UID: \"9d4974fb-9d3c-4e2d-ac54-7c3d4b4b3f89\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qznqm" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.654632 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/adc0635a-397c-48ba-b103-a8d6e62ae535-srv-cert\") pod \"olm-operator-6b444d44fb-5j88f\" (UID: \"adc0635a-397c-48ba-b103-a8d6e62ae535\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5j88f" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.654649 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88cbx\" (UniqueName: \"kubernetes.io/projected/3a41af56-8a6d-49ba-8b85-1c546d49c198-kube-api-access-88cbx\") pod \"dns-default-nvt5z\" (UID: \"3a41af56-8a6d-49ba-8b85-1c546d49c198\") " pod="openshift-dns/dns-default-nvt5z" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.654689 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvwhg\" (UniqueName: \"kubernetes.io/projected/7336d2fb-f015-4e22-8419-fdefc49f6137-kube-api-access-dvwhg\") pod \"catalog-operator-68c6474976-8kfz8\" (UID: \"7336d2fb-f015-4e22-8419-fdefc49f6137\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8kfz8" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.654724 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x55xr\" (UniqueName: \"kubernetes.io/projected/a8b41c08-4e89-4f08-8d45-256c306fbea0-kube-api-access-x55xr\") pod \"marketplace-operator-79b997595-tfc7g\" (UID: \"a8b41c08-4e89-4f08-8d45-256c306fbea0\") " pod="openshift-marketplace/marketplace-operator-79b997595-tfc7g" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.654748 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a8b41c08-4e89-4f08-8d45-256c306fbea0-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tfc7g\" (UID: \"a8b41c08-4e89-4f08-8d45-256c306fbea0\") " pod="openshift-marketplace/marketplace-operator-79b997595-tfc7g" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.654763 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b16d0ad2-0b45-45fe-bc2c-215697447010-serving-cert\") pod \"service-ca-operator-777779d784-ktnrj\" (UID: \"b16d0ad2-0b45-45fe-bc2c-215697447010\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ktnrj" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.654851 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/80a23793-4622-4288-a5c0-de5199e1e767-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-xg5l9\" (UID: \"80a23793-4622-4288-a5c0-de5199e1e767\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xg5l9" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.654873 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/912dc882-2c54-4a47-b26b-6f3a63f9f33f-tmpfs\") pod \"packageserver-d55dfcdfc-rtd5w\" (UID: \"912dc882-2c54-4a47-b26b-6f3a63f9f33f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rtd5w" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.654892 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fw6b\" (UniqueName: \"kubernetes.io/projected/bf0e3a33-4470-430f-8ebc-71d52b649684-kube-api-access-4fw6b\") pod \"migrator-59844c95c7-p559m\" (UID: \"bf0e3a33-4470-430f-8ebc-71d52b649684\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-p559m" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.654906 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stj62\" (UniqueName: \"kubernetes.io/projected/80a23793-4622-4288-a5c0-de5199e1e767-kube-api-access-stj62\") pod \"package-server-manager-789f6589d5-xg5l9\" (UID: \"80a23793-4622-4288-a5c0-de5199e1e767\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xg5l9" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.654922 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/093665e3-6dc0-4c51-a0f3-71363eef8367-stats-auth\") pod \"router-default-5444994796-85fhx\" (UID: \"093665e3-6dc0-4c51-a0f3-71363eef8367\") " pod="openshift-ingress/router-default-5444994796-85fhx" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.654956 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7336d2fb-f015-4e22-8419-fdefc49f6137-srv-cert\") pod \"catalog-operator-68c6474976-8kfz8\" (UID: \"7336d2fb-f015-4e22-8419-fdefc49f6137\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8kfz8" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.654975 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzfch\" (UniqueName: \"kubernetes.io/projected/b16d0ad2-0b45-45fe-bc2c-215697447010-kube-api-access-mzfch\") pod \"service-ca-operator-777779d784-ktnrj\" (UID: \"b16d0ad2-0b45-45fe-bc2c-215697447010\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ktnrj" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.655021 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5057f820-3d60-43e3-9cb8-ceac9650621d-config\") pod \"kube-apiserver-operator-766d6c64bb-kfp77\" (UID: \"5057f820-3d60-43e3-9cb8-ceac9650621d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kfp77" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.655036 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/77153235-6efa-4db2-a4a3-5153d9ba3568-certs\") pod \"machine-config-server-4v795\" (UID: \"77153235-6efa-4db2-a4a3-5153d9ba3568\") " pod="openshift-machine-config-operator/machine-config-server-4v795" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.655063 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/cd220c67-e881-4180-b1c4-4064be0794ca-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-k77jn\" (UID: \"cd220c67-e881-4180-b1c4-4064be0794ca\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-k77jn" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.655081 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/adc0635a-397c-48ba-b103-a8d6e62ae535-profile-collector-cert\") pod \"olm-operator-6b444d44fb-5j88f\" (UID: \"adc0635a-397c-48ba-b103-a8d6e62ae535\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5j88f" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.655100 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/912dc882-2c54-4a47-b26b-6f3a63f9f33f-apiservice-cert\") pod \"packageserver-d55dfcdfc-rtd5w\" (UID: \"912dc882-2c54-4a47-b26b-6f3a63f9f33f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rtd5w" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.655118 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66ddt\" (UniqueName: \"kubernetes.io/projected/7f11b3c3-6380-4f4d-9d03-622e66e33bc8-kube-api-access-66ddt\") pod \"collect-profiles-29325780-9b22t\" (UID: \"7f11b3c3-6380-4f4d-9d03-622e66e33bc8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325780-9b22t" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.655135 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4ba39c3a-d31f-4f1d-807d-c9fcf6defa34-proxy-tls\") pod \"machine-config-operator-74547568cd-6prnz\" (UID: \"4ba39c3a-d31f-4f1d-807d-c9fcf6defa34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6prnz" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.655161 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4ba39c3a-d31f-4f1d-807d-c9fcf6defa34-auth-proxy-config\") pod \"machine-config-operator-74547568cd-6prnz\" (UID: \"4ba39c3a-d31f-4f1d-807d-c9fcf6defa34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6prnz" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.655177 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkntl\" (UniqueName: \"kubernetes.io/projected/c151b75e-6548-4d28-851e-d8f4eac68183-kube-api-access-bkntl\") pod \"service-ca-9c57cc56f-vtmjj\" (UID: \"c151b75e-6548-4d28-851e-d8f4eac68183\") " pod="openshift-service-ca/service-ca-9c57cc56f-vtmjj" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.655226 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/9d4974fb-9d3c-4e2d-ac54-7c3d4b4b3f89-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-qznqm\" (UID: \"9d4974fb-9d3c-4e2d-ac54-7c3d4b4b3f89\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qznqm" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.655243 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9x4m2\" (UniqueName: \"kubernetes.io/projected/35542dc5-96db-4eca-a2c2-85ebf65ea31f-kube-api-access-9x4m2\") pod \"csi-hostpathplugin-wn8jl\" (UID: \"35542dc5-96db-4eca-a2c2-85ebf65ea31f\") " pod="hostpath-provisioner/csi-hostpathplugin-wn8jl" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.655240 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/35542dc5-96db-4eca-a2c2-85ebf65ea31f-socket-dir\") pod \"csi-hostpathplugin-wn8jl\" (UID: \"35542dc5-96db-4eca-a2c2-85ebf65ea31f\") " pod="hostpath-provisioner/csi-hostpathplugin-wn8jl" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.655613 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vztvs\" (UniqueName: \"kubernetes.io/projected/cd220c67-e881-4180-b1c4-4064be0794ca-kube-api-access-vztvs\") pod \"multus-admission-controller-857f4d67dd-k77jn\" (UID: \"cd220c67-e881-4180-b1c4-4064be0794ca\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-k77jn" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.655654 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/530fb6b6-9c36-4d13-8aa9-903d7080082f-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-krpzs\" (UID: \"530fb6b6-9c36-4d13-8aa9-903d7080082f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-krpzs" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.655674 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5057f820-3d60-43e3-9cb8-ceac9650621d-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-kfp77\" (UID: \"5057f820-3d60-43e3-9cb8-ceac9650621d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kfp77" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.655694 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbhrp\" (UniqueName: \"kubernetes.io/projected/4ba39c3a-d31f-4f1d-807d-c9fcf6defa34-kube-api-access-xbhrp\") pod \"machine-config-operator-74547568cd-6prnz\" (UID: \"4ba39c3a-d31f-4f1d-807d-c9fcf6defa34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6prnz" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.655726 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2556513e-2080-4ed1-99cc-bb1e15a75a8e-cert\") pod \"ingress-canary-lgfzr\" (UID: \"2556513e-2080-4ed1-99cc-bb1e15a75a8e\") " pod="openshift-ingress-canary/ingress-canary-lgfzr" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.655755 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/35542dc5-96db-4eca-a2c2-85ebf65ea31f-registration-dir\") pod \"csi-hostpathplugin-wn8jl\" (UID: \"35542dc5-96db-4eca-a2c2-85ebf65ea31f\") " pod="hostpath-provisioner/csi-hostpathplugin-wn8jl" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.655774 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xl64r\" (UniqueName: \"kubernetes.io/projected/093665e3-6dc0-4c51-a0f3-71363eef8367-kube-api-access-xl64r\") pod \"router-default-5444994796-85fhx\" (UID: \"093665e3-6dc0-4c51-a0f3-71363eef8367\") " pod="openshift-ingress/router-default-5444994796-85fhx" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.655790 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnlxv\" (UniqueName: \"kubernetes.io/projected/adc0635a-397c-48ba-b103-a8d6e62ae535-kube-api-access-jnlxv\") pod \"olm-operator-6b444d44fb-5j88f\" (UID: \"adc0635a-397c-48ba-b103-a8d6e62ae535\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5j88f" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.655807 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7336d2fb-f015-4e22-8419-fdefc49f6137-profile-collector-cert\") pod \"catalog-operator-68c6474976-8kfz8\" (UID: \"7336d2fb-f015-4e22-8419-fdefc49f6137\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8kfz8" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.655832 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r679r\" (UniqueName: \"kubernetes.io/projected/2556513e-2080-4ed1-99cc-bb1e15a75a8e-kube-api-access-r679r\") pod \"ingress-canary-lgfzr\" (UID: \"2556513e-2080-4ed1-99cc-bb1e15a75a8e\") " pod="openshift-ingress-canary/ingress-canary-lgfzr" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.665939 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/adc0635a-397c-48ba-b103-a8d6e62ae535-srv-cert\") pod \"olm-operator-6b444d44fb-5j88f\" (UID: \"adc0635a-397c-48ba-b103-a8d6e62ae535\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5j88f" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.668346 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/093665e3-6dc0-4c51-a0f3-71363eef8367-default-certificate\") pod \"router-default-5444994796-85fhx\" (UID: \"093665e3-6dc0-4c51-a0f3-71363eef8367\") " pod="openshift-ingress/router-default-5444994796-85fhx" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.669333 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b16d0ad2-0b45-45fe-bc2c-215697447010-config\") pod \"service-ca-operator-777779d784-ktnrj\" (UID: \"b16d0ad2-0b45-45fe-bc2c-215697447010\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ktnrj" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.669563 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a8b41c08-4e89-4f08-8d45-256c306fbea0-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tfc7g\" (UID: \"a8b41c08-4e89-4f08-8d45-256c306fbea0\") " pod="openshift-marketplace/marketplace-operator-79b997595-tfc7g" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.669970 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8edba022-cd27-43b2-87e3-a0e92562c450-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-rlbvb\" (UID: \"8edba022-cd27-43b2-87e3-a0e92562c450\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rlbvb" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.670346 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/530fb6b6-9c36-4d13-8aa9-903d7080082f-proxy-tls\") pod \"machine-config-controller-84d6567774-krpzs\" (UID: \"530fb6b6-9c36-4d13-8aa9-903d7080082f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-krpzs" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.670376 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/35542dc5-96db-4eca-a2c2-85ebf65ea31f-plugins-dir\") pod \"csi-hostpathplugin-wn8jl\" (UID: \"35542dc5-96db-4eca-a2c2-85ebf65ea31f\") " pod="hostpath-provisioner/csi-hostpathplugin-wn8jl" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.670400 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7f11b3c3-6380-4f4d-9d03-622e66e33bc8-config-volume\") pod \"collect-profiles-29325780-9b22t\" (UID: \"7f11b3c3-6380-4f4d-9d03-622e66e33bc8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325780-9b22t" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.670447 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ks892\" (UniqueName: \"kubernetes.io/projected/912dc882-2c54-4a47-b26b-6f3a63f9f33f-kube-api-access-ks892\") pod \"packageserver-d55dfcdfc-rtd5w\" (UID: \"912dc882-2c54-4a47-b26b-6f3a63f9f33f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rtd5w" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.670473 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/35542dc5-96db-4eca-a2c2-85ebf65ea31f-mountpoint-dir\") pod \"csi-hostpathplugin-wn8jl\" (UID: \"35542dc5-96db-4eca-a2c2-85ebf65ea31f\") " pod="hostpath-provisioner/csi-hostpathplugin-wn8jl" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.670501 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v42w5\" (UniqueName: \"kubernetes.io/projected/77153235-6efa-4db2-a4a3-5153d9ba3568-kube-api-access-v42w5\") pod \"machine-config-server-4v795\" (UID: \"77153235-6efa-4db2-a4a3-5153d9ba3568\") " pod="openshift-machine-config-operator/machine-config-server-4v795" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.670544 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5057f820-3d60-43e3-9cb8-ceac9650621d-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-kfp77\" (UID: \"5057f820-3d60-43e3-9cb8-ceac9650621d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kfp77" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.670593 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.670622 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/c151b75e-6548-4d28-851e-d8f4eac68183-signing-cabundle\") pod \"service-ca-9c57cc56f-vtmjj\" (UID: \"c151b75e-6548-4d28-851e-d8f4eac68183\") " pod="openshift-service-ca/service-ca-9c57cc56f-vtmjj" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.670688 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7f11b3c3-6380-4f4d-9d03-622e66e33bc8-secret-volume\") pod \"collect-profiles-29325780-9b22t\" (UID: \"7f11b3c3-6380-4f4d-9d03-622e66e33bc8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325780-9b22t" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.670710 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4ba39c3a-d31f-4f1d-807d-c9fcf6defa34-images\") pod \"machine-config-operator-74547568cd-6prnz\" (UID: \"4ba39c3a-d31f-4f1d-807d-c9fcf6defa34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6prnz" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.670733 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/093665e3-6dc0-4c51-a0f3-71363eef8367-service-ca-bundle\") pod \"router-default-5444994796-85fhx\" (UID: \"093665e3-6dc0-4c51-a0f3-71363eef8367\") " pod="openshift-ingress/router-default-5444994796-85fhx" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.670752 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3a41af56-8a6d-49ba-8b85-1c546d49c198-config-volume\") pod \"dns-default-nvt5z\" (UID: \"3a41af56-8a6d-49ba-8b85-1c546d49c198\") " pod="openshift-dns/dns-default-nvt5z" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.670777 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/77153235-6efa-4db2-a4a3-5153d9ba3568-node-bootstrap-token\") pod \"machine-config-server-4v795\" (UID: \"77153235-6efa-4db2-a4a3-5153d9ba3568\") " pod="openshift-machine-config-operator/machine-config-server-4v795" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.671061 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pkdzv" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.672520 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/c151b75e-6548-4d28-851e-d8f4eac68183-signing-key\") pod \"service-ca-9c57cc56f-vtmjj\" (UID: \"c151b75e-6548-4d28-851e-d8f4eac68183\") " pod="openshift-service-ca/service-ca-9c57cc56f-vtmjj" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.676259 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/35542dc5-96db-4eca-a2c2-85ebf65ea31f-registration-dir\") pod \"csi-hostpathplugin-wn8jl\" (UID: \"35542dc5-96db-4eca-a2c2-85ebf65ea31f\") " pod="hostpath-provisioner/csi-hostpathplugin-wn8jl" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.679424 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5057f820-3d60-43e3-9cb8-ceac9650621d-config\") pod \"kube-apiserver-operator-766d6c64bb-kfp77\" (UID: \"5057f820-3d60-43e3-9cb8-ceac9650621d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kfp77" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.680125 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/530fb6b6-9c36-4d13-8aa9-903d7080082f-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-krpzs\" (UID: \"530fb6b6-9c36-4d13-8aa9-903d7080082f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-krpzs" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.680803 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2556513e-2080-4ed1-99cc-bb1e15a75a8e-cert\") pod \"ingress-canary-lgfzr\" (UID: \"2556513e-2080-4ed1-99cc-bb1e15a75a8e\") " pod="openshift-ingress-canary/ingress-canary-lgfzr" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.681187 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3a41af56-8a6d-49ba-8b85-1c546d49c198-metrics-tls\") pod \"dns-default-nvt5z\" (UID: \"3a41af56-8a6d-49ba-8b85-1c546d49c198\") " pod="openshift-dns/dns-default-nvt5z" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.685820 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5057f820-3d60-43e3-9cb8-ceac9650621d-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-kfp77\" (UID: \"5057f820-3d60-43e3-9cb8-ceac9650621d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kfp77" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.686508 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7336d2fb-f015-4e22-8419-fdefc49f6137-profile-collector-cert\") pod \"catalog-operator-68c6474976-8kfz8\" (UID: \"7336d2fb-f015-4e22-8419-fdefc49f6137\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8kfz8" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.689046 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/093665e3-6dc0-4c51-a0f3-71363eef8367-metrics-certs\") pod \"router-default-5444994796-85fhx\" (UID: \"093665e3-6dc0-4c51-a0f3-71363eef8367\") " pod="openshift-ingress/router-default-5444994796-85fhx" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.689109 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/35542dc5-96db-4eca-a2c2-85ebf65ea31f-plugins-dir\") pod \"csi-hostpathplugin-wn8jl\" (UID: \"35542dc5-96db-4eca-a2c2-85ebf65ea31f\") " pod="hostpath-provisioner/csi-hostpathplugin-wn8jl" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.689216 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7f11b3c3-6380-4f4d-9d03-622e66e33bc8-config-volume\") pod \"collect-profiles-29325780-9b22t\" (UID: \"7f11b3c3-6380-4f4d-9d03-622e66e33bc8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325780-9b22t" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.689256 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4ba39c3a-d31f-4f1d-807d-c9fcf6defa34-auth-proxy-config\") pod \"machine-config-operator-74547568cd-6prnz\" (UID: \"4ba39c3a-d31f-4f1d-807d-c9fcf6defa34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6prnz" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.689231 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/912dc882-2c54-4a47-b26b-6f3a63f9f33f-webhook-cert\") pod \"packageserver-d55dfcdfc-rtd5w\" (UID: \"912dc882-2c54-4a47-b26b-6f3a63f9f33f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rtd5w" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.690682 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/35542dc5-96db-4eca-a2c2-85ebf65ea31f-mountpoint-dir\") pod \"csi-hostpathplugin-wn8jl\" (UID: \"35542dc5-96db-4eca-a2c2-85ebf65ea31f\") " pod="hostpath-provisioner/csi-hostpathplugin-wn8jl" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.691225 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/912dc882-2c54-4a47-b26b-6f3a63f9f33f-tmpfs\") pod \"packageserver-d55dfcdfc-rtd5w\" (UID: \"912dc882-2c54-4a47-b26b-6f3a63f9f33f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rtd5w" Oct 04 03:12:55 crc kubenswrapper[4742]: E1004 03:12:55.692712 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:12:56.192690032 +0000 UTC m=+142.125027208 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.695008 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/c151b75e-6548-4d28-851e-d8f4eac68183-signing-cabundle\") pod \"service-ca-9c57cc56f-vtmjj\" (UID: \"c151b75e-6548-4d28-851e-d8f4eac68183\") " pod="openshift-service-ca/service-ca-9c57cc56f-vtmjj" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.695567 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/530fb6b6-9c36-4d13-8aa9-903d7080082f-proxy-tls\") pod \"machine-config-controller-84d6567774-krpzs\" (UID: \"530fb6b6-9c36-4d13-8aa9-903d7080082f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-krpzs" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.695725 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/093665e3-6dc0-4c51-a0f3-71363eef8367-service-ca-bundle\") pod \"router-default-5444994796-85fhx\" (UID: \"093665e3-6dc0-4c51-a0f3-71363eef8367\") " pod="openshift-ingress/router-default-5444994796-85fhx" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.697641 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3a41af56-8a6d-49ba-8b85-1c546d49c198-config-volume\") pod \"dns-default-nvt5z\" (UID: \"3a41af56-8a6d-49ba-8b85-1c546d49c198\") " pod="openshift-dns/dns-default-nvt5z" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.698142 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/80a23793-4622-4288-a5c0-de5199e1e767-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-xg5l9\" (UID: \"80a23793-4622-4288-a5c0-de5199e1e767\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xg5l9" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.699348 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4ba39c3a-d31f-4f1d-807d-c9fcf6defa34-images\") pod \"machine-config-operator-74547568cd-6prnz\" (UID: \"4ba39c3a-d31f-4f1d-807d-c9fcf6defa34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6prnz" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.702770 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4892c4de-41a2-4c58-bb59-36a6c584da2f-bound-sa-token\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.704734 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/912dc882-2c54-4a47-b26b-6f3a63f9f33f-apiservice-cert\") pod \"packageserver-d55dfcdfc-rtd5w\" (UID: \"912dc882-2c54-4a47-b26b-6f3a63f9f33f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rtd5w" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.723648 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/adc0635a-397c-48ba-b103-a8d6e62ae535-profile-collector-cert\") pod \"olm-operator-6b444d44fb-5j88f\" (UID: \"adc0635a-397c-48ba-b103-a8d6e62ae535\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5j88f" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.737541 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6nmn2" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.742239 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/9d4974fb-9d3c-4e2d-ac54-7c3d4b4b3f89-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-qznqm\" (UID: \"9d4974fb-9d3c-4e2d-ac54-7c3d4b4b3f89\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qznqm" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.743837 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/cd220c67-e881-4180-b1c4-4064be0794ca-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-k77jn\" (UID: \"cd220c67-e881-4180-b1c4-4064be0794ca\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-k77jn" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.744325 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.753184 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rlbvb" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.757691 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7f11b3c3-6380-4f4d-9d03-622e66e33bc8-secret-volume\") pod \"collect-profiles-29325780-9b22t\" (UID: \"7f11b3c3-6380-4f4d-9d03-622e66e33bc8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325780-9b22t" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.763407 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9s6j\" (UniqueName: \"kubernetes.io/projected/bb08fe29-5abf-43af-a622-f1db8e955e55-kube-api-access-h9s6j\") pod \"openshift-config-operator-7777fb866f-8w75v\" (UID: \"bb08fe29-5abf-43af-a622-f1db8e955e55\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8w75v" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.767951 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/77153235-6efa-4db2-a4a3-5153d9ba3568-certs\") pod \"machine-config-server-4v795\" (UID: \"77153235-6efa-4db2-a4a3-5153d9ba3568\") " pod="openshift-machine-config-operator/machine-config-server-4v795" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.768442 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7336d2fb-f015-4e22-8419-fdefc49f6137-srv-cert\") pod \"catalog-operator-68c6474976-8kfz8\" (UID: \"7336d2fb-f015-4e22-8419-fdefc49f6137\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8kfz8" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.772680 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b16d0ad2-0b45-45fe-bc2c-215697447010-serving-cert\") pod \"service-ca-operator-777779d784-ktnrj\" (UID: \"b16d0ad2-0b45-45fe-bc2c-215697447010\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ktnrj" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.772760 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/093665e3-6dc0-4c51-a0f3-71363eef8367-stats-auth\") pod \"router-default-5444994796-85fhx\" (UID: \"093665e3-6dc0-4c51-a0f3-71363eef8367\") " pod="openshift-ingress/router-default-5444994796-85fhx" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.773206 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/77153235-6efa-4db2-a4a3-5153d9ba3568-node-bootstrap-token\") pod \"machine-config-server-4v795\" (UID: \"77153235-6efa-4db2-a4a3-5153d9ba3568\") " pod="openshift-machine-config-operator/machine-config-server-4v795" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.780870 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnxpt\" (UniqueName: \"kubernetes.io/projected/528adbb9-98ba-45d3-ba3a-2fe715e65cf6-kube-api-access-rnxpt\") pod \"dns-operator-744455d44c-cms9s\" (UID: \"528adbb9-98ba-45d3-ba3a-2fe715e65cf6\") " pod="openshift-dns-operator/dns-operator-744455d44c-cms9s" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.781098 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-v92dt" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.786868 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7940c6dd-4863-4c92-ad0e-27abb961da4a-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-dmd8c\" (UID: \"7940c6dd-4863-4c92-ad0e-27abb961da4a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-dmd8c" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.787004 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk46r\" (UniqueName: \"kubernetes.io/projected/4892c4de-41a2-4c58-bb59-36a6c584da2f-kube-api-access-tk46r\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.791821 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:12:55 crc kubenswrapper[4742]: E1004 03:12:55.792212 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:12:56.292188358 +0000 UTC m=+142.224525534 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.792403 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:55 crc kubenswrapper[4742]: E1004 03:12:55.793042 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:12:56.293030391 +0000 UTC m=+142.225367567 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.793870 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4ba39c3a-d31f-4f1d-807d-c9fcf6defa34-proxy-tls\") pod \"machine-config-operator-74547568cd-6prnz\" (UID: \"4ba39c3a-d31f-4f1d-807d-c9fcf6defa34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6prnz" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.796610 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a8b41c08-4e89-4f08-8d45-256c306fbea0-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tfc7g\" (UID: \"a8b41c08-4e89-4f08-8d45-256c306fbea0\") " pod="openshift-marketplace/marketplace-operator-79b997595-tfc7g" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.796877 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-dmd8c" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.811020 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blwh2\" (UniqueName: \"kubernetes.io/projected/530fb6b6-9c36-4d13-8aa9-903d7080082f-kube-api-access-blwh2\") pod \"machine-config-controller-84d6567774-krpzs\" (UID: \"530fb6b6-9c36-4d13-8aa9-903d7080082f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-krpzs" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.820594 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88cbx\" (UniqueName: \"kubernetes.io/projected/3a41af56-8a6d-49ba-8b85-1c546d49c198-kube-api-access-88cbx\") pod \"dns-default-nvt5z\" (UID: \"3a41af56-8a6d-49ba-8b85-1c546d49c198\") " pod="openshift-dns/dns-default-nvt5z" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.829447 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ncmp2" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.867399 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mspvb"] Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.868668 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpsq6\" (UniqueName: \"kubernetes.io/projected/9d4974fb-9d3c-4e2d-ac54-7c3d4b4b3f89-kube-api-access-wpsq6\") pod \"cluster-samples-operator-665b6dd947-qznqm\" (UID: \"9d4974fb-9d3c-4e2d-ac54-7c3d4b4b3f89\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qznqm" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.871685 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66ddt\" (UniqueName: \"kubernetes.io/projected/7f11b3c3-6380-4f4d-9d03-622e66e33bc8-kube-api-access-66ddt\") pod \"collect-profiles-29325780-9b22t\" (UID: \"7f11b3c3-6380-4f4d-9d03-622e66e33bc8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325780-9b22t" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.880298 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325780-9b22t" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.882075 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-ctkcp"] Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.888766 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-h22jn"] Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.888944 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stj62\" (UniqueName: \"kubernetes.io/projected/80a23793-4622-4288-a5c0-de5199e1e767-kube-api-access-stj62\") pod \"package-server-manager-789f6589d5-xg5l9\" (UID: \"80a23793-4622-4288-a5c0-de5199e1e767\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xg5l9" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.893848 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:12:55 crc kubenswrapper[4742]: E1004 03:12:55.894003 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:12:56.393975656 +0000 UTC m=+142.326312832 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.894346 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:55 crc kubenswrapper[4742]: E1004 03:12:55.894766 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:12:56.394748926 +0000 UTC m=+142.327086102 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.896613 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qznqm" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.903718 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fw6b\" (UniqueName: \"kubernetes.io/projected/bf0e3a33-4470-430f-8ebc-71d52b649684-kube-api-access-4fw6b\") pod \"migrator-59844c95c7-p559m\" (UID: \"bf0e3a33-4470-430f-8ebc-71d52b649684\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-p559m" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.912555 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-2pdbs"] Oct 04 03:12:55 crc kubenswrapper[4742]: W1004 03:12:55.918174 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0e5511f2_7587_4638_bd0c_af91c57a0163.slice/crio-d975d272f7dbbfe0ebee90accbb3ccbf015123ea200150c340b8da79b572569e WatchSource:0}: Error finding container d975d272f7dbbfe0ebee90accbb3ccbf015123ea200150c340b8da79b572569e: Status 404 returned error can't find the container with id d975d272f7dbbfe0ebee90accbb3ccbf015123ea200150c340b8da79b572569e Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.918586 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xg5l9" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.926032 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvwhg\" (UniqueName: \"kubernetes.io/projected/7336d2fb-f015-4e22-8419-fdefc49f6137-kube-api-access-dvwhg\") pod \"catalog-operator-68c6474976-8kfz8\" (UID: \"7336d2fb-f015-4e22-8419-fdefc49f6137\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8kfz8" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.935257 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5057f820-3d60-43e3-9cb8-ceac9650621d-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-kfp77\" (UID: \"5057f820-3d60-43e3-9cb8-ceac9650621d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kfp77" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.939841 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-krpzs" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.957458 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbhrp\" (UniqueName: \"kubernetes.io/projected/4ba39c3a-d31f-4f1d-807d-c9fcf6defa34-kube-api-access-xbhrp\") pod \"machine-config-operator-74547568cd-6prnz\" (UID: \"4ba39c3a-d31f-4f1d-807d-c9fcf6defa34\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6prnz" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.981754 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x55xr\" (UniqueName: \"kubernetes.io/projected/a8b41c08-4e89-4f08-8d45-256c306fbea0-kube-api-access-x55xr\") pod \"marketplace-operator-79b997595-tfc7g\" (UID: \"a8b41c08-4e89-4f08-8d45-256c306fbea0\") " pod="openshift-marketplace/marketplace-operator-79b997595-tfc7g" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.989651 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-nvt5z" Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.994874 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:12:55 crc kubenswrapper[4742]: E1004 03:12:55.995108 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:12:56.495066084 +0000 UTC m=+142.427403280 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:55 crc kubenswrapper[4742]: I1004 03:12:55.995574 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:55 crc kubenswrapper[4742]: E1004 03:12:55.998059 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:12:56.498029504 +0000 UTC m=+142.430366850 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.002874 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xl64r\" (UniqueName: \"kubernetes.io/projected/093665e3-6dc0-4c51-a0f3-71363eef8367-kube-api-access-xl64r\") pod \"router-default-5444994796-85fhx\" (UID: \"093665e3-6dc0-4c51-a0f3-71363eef8367\") " pod="openshift-ingress/router-default-5444994796-85fhx" Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.021913 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnlxv\" (UniqueName: \"kubernetes.io/projected/adc0635a-397c-48ba-b103-a8d6e62ae535-kube-api-access-jnlxv\") pod \"olm-operator-6b444d44fb-5j88f\" (UID: \"adc0635a-397c-48ba-b103-a8d6e62ae535\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5j88f" Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.046525 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzfch\" (UniqueName: \"kubernetes.io/projected/b16d0ad2-0b45-45fe-bc2c-215697447010-kube-api-access-mzfch\") pod \"service-ca-operator-777779d784-ktnrj\" (UID: \"b16d0ad2-0b45-45fe-bc2c-215697447010\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ktnrj" Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.047125 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mgh2f"] Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.057740 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8w75v" Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.058173 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r679r\" (UniqueName: \"kubernetes.io/projected/2556513e-2080-4ed1-99cc-bb1e15a75a8e-kube-api-access-r679r\") pod \"ingress-canary-lgfzr\" (UID: \"2556513e-2080-4ed1-99cc-bb1e15a75a8e\") " pod="openshift-ingress-canary/ingress-canary-lgfzr" Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.072616 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-cms9s" Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.078362 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v42w5\" (UniqueName: \"kubernetes.io/projected/77153235-6efa-4db2-a4a3-5153d9ba3568-kube-api-access-v42w5\") pod \"machine-config-server-4v795\" (UID: \"77153235-6efa-4db2-a4a3-5153d9ba3568\") " pod="openshift-machine-config-operator/machine-config-server-4v795" Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.098754 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:12:56 crc kubenswrapper[4742]: E1004 03:12:56.099250 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:12:56.599229105 +0000 UTC m=+142.531566281 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.110543 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ks892\" (UniqueName: \"kubernetes.io/projected/912dc882-2c54-4a47-b26b-6f3a63f9f33f-kube-api-access-ks892\") pod \"packageserver-d55dfcdfc-rtd5w\" (UID: \"912dc882-2c54-4a47-b26b-6f3a63f9f33f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rtd5w" Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.122620 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9x4m2\" (UniqueName: \"kubernetes.io/projected/35542dc5-96db-4eca-a2c2-85ebf65ea31f-kube-api-access-9x4m2\") pod \"csi-hostpathplugin-wn8jl\" (UID: \"35542dc5-96db-4eca-a2c2-85ebf65ea31f\") " pod="hostpath-provisioner/csi-hostpathplugin-wn8jl" Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.136072 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-p559m" Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.142324 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkntl\" (UniqueName: \"kubernetes.io/projected/c151b75e-6548-4d28-851e-d8f4eac68183-kube-api-access-bkntl\") pod \"service-ca-9c57cc56f-vtmjj\" (UID: \"c151b75e-6548-4d28-851e-d8f4eac68183\") " pod="openshift-service-ca/service-ca-9c57cc56f-vtmjj" Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.143454 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-85fhx" Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.152690 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kfp77" Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.158739 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-tfc7g" Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.161167 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t7pnt" Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.161959 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vztvs\" (UniqueName: \"kubernetes.io/projected/cd220c67-e881-4180-b1c4-4064be0794ca-kube-api-access-vztvs\") pod \"multus-admission-controller-857f4d67dd-k77jn\" (UID: \"cd220c67-e881-4180-b1c4-4064be0794ca\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-k77jn" Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.166554 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6prnz" Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.174058 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rtd5w" Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.179945 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nsspd"] Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.187797 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-4v795" Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.196377 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-fvr9s"] Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.202840 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:56 crc kubenswrapper[4742]: E1004 03:12:56.203198 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:12:56.703185701 +0000 UTC m=+142.635522877 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.203266 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5j88f" Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.210064 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8kfz8" Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.234085 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-ktnrj" Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.247904 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-k77jn" Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.255579 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-vtmjj" Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.264175 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-lgfzr" Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.293657 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-wn8jl" Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.303860 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:12:56 crc kubenswrapper[4742]: E1004 03:12:56.306964 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:12:56.806927051 +0000 UTC m=+142.739264227 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.307942 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:56 crc kubenswrapper[4742]: E1004 03:12:56.308312 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:12:56.808297807 +0000 UTC m=+142.740634973 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.311786 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-mfp6b"] Oct 04 03:12:56 crc kubenswrapper[4742]: W1004 03:12:56.323679 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc975f343_d9c0_4145_b0b8_bdc11dbd1d66.slice/crio-7f1c0759482b8a49cc643c1c0b4d6644ec99e4fd07c98b6359a3dbccb510e868 WatchSource:0}: Error finding container 7f1c0759482b8a49cc643c1c0b4d6644ec99e4fd07c98b6359a3dbccb510e868: Status 404 returned error can't find the container with id 7f1c0759482b8a49cc643c1c0b4d6644ec99e4fd07c98b6359a3dbccb510e868 Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.416131 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:12:56 crc kubenswrapper[4742]: E1004 03:12:56.416769 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:12:56.916734643 +0000 UTC m=+142.849071819 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.518306 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:56 crc kubenswrapper[4742]: E1004 03:12:56.518750 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:12:57.018730956 +0000 UTC m=+142.951068302 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:56 crc kubenswrapper[4742]: W1004 03:12:56.557255 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3b81c30_6a85_430c_a915_6322d20953ac.slice/crio-5f42521310edd9aabdfad982fc7f72e7863334a93eb6a18148df4c44465b555d WatchSource:0}: Error finding container 5f42521310edd9aabdfad982fc7f72e7863334a93eb6a18148df4c44465b555d: Status 404 returned error can't find the container with id 5f42521310edd9aabdfad982fc7f72e7863334a93eb6a18148df4c44465b555d Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.594342 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-dmd8c"] Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.602710 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q"] Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.602791 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-v92dt"] Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.620496 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:12:56 crc kubenswrapper[4742]: E1004 03:12:56.620724 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:12:57.120670588 +0000 UTC m=+143.053007764 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.621442 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:56 crc kubenswrapper[4742]: W1004 03:12:56.621678 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd33e79f4_4871_4e03_ac04_2000ceb3416d.slice/crio-cb42f04d6bf27a3044735b0f0b83a1f2339e52a4597a1bc58a2acc0d249b3441 WatchSource:0}: Error finding container cb42f04d6bf27a3044735b0f0b83a1f2339e52a4597a1bc58a2acc0d249b3441: Status 404 returned error can't find the container with id cb42f04d6bf27a3044735b0f0b83a1f2339e52a4597a1bc58a2acc0d249b3441 Oct 04 03:12:56 crc kubenswrapper[4742]: E1004 03:12:56.621877 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:12:57.12186593 +0000 UTC m=+143.054203106 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.624829 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ncmp2"] Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.627078 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rlbvb"] Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.638046 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mfp6b" event={"ID":"e3b81c30-6a85-430c-a915-6322d20953ac","Type":"ContainerStarted","Data":"5f42521310edd9aabdfad982fc7f72e7863334a93eb6a18148df4c44465b555d"} Oct 04 03:12:56 crc kubenswrapper[4742]: W1004 03:12:56.638875 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7940c6dd_4863_4c92_ad0e_27abb961da4a.slice/crio-f7d70a6ae4e354e2b217b767ceef38b9c2628aa10d6ef40a57e7d068a28c1887 WatchSource:0}: Error finding container f7d70a6ae4e354e2b217b767ceef38b9c2628aa10d6ef40a57e7d068a28c1887: Status 404 returned error can't find the container with id f7d70a6ae4e354e2b217b767ceef38b9c2628aa10d6ef40a57e7d068a28c1887 Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.639688 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mgh2f" event={"ID":"c975f343-d9c0-4145-b0b8-bdc11dbd1d66","Type":"ContainerStarted","Data":"7f1c0759482b8a49cc643c1c0b4d6644ec99e4fd07c98b6359a3dbccb510e868"} Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.641244 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-fvr9s" event={"ID":"09503673-286e-48af-b18b-37bff87f8019","Type":"ContainerStarted","Data":"39d5b7d406a8536ec22dc36acaa35ad61d6b7f3aa8c72420287afef8a508e69b"} Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.642494 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" event={"ID":"56722804-fa7c-4639-a4d3-5085cd965f8b","Type":"ContainerStarted","Data":"b9ca38ac05caade3628be76b9f62478ef0999ffad533cf94b7cc6ff140a9c7bb"} Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.657478 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pkdzv" event={"ID":"da95bbc1-a1c0-428f-9884-cda886a8da3c","Type":"ContainerStarted","Data":"457b035caccfe9d4b7a0d4cd5897d0931d9b7ef660e093355a50c565480a02aa"} Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.662190 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nsspd" event={"ID":"8725e189-8b15-4c13-ae9b-f1230476e572","Type":"ContainerStarted","Data":"43d6717c37d60edf2d65fbaa44d665b027359339baa7f2476dabf5a352401143"} Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.670337 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-2pdbs" event={"ID":"07f1cf4a-506e-469a-a8e5-0adc68a5ab81","Type":"ContainerStarted","Data":"5c6b771fe411ff11924b832ec902b0545e11cfa85365b8c5a1a550570078ced6"} Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.672482 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ctkcp" event={"ID":"5ba77307-01a4-43a3-8888-3da1c700ea2e","Type":"ContainerStarted","Data":"4d32e817a4ba418c256fc7cbdea18289424919779de9b1954ac0fd0bb39eded2"} Oct 04 03:12:56 crc kubenswrapper[4742]: W1004 03:12:56.674585 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d53af45_9fb2_4d36_8c21_aa167c8ed70f.slice/crio-a43cff83e2e19f182e5b5e1e17e35d28c0cea02c7fff97a3c46f427a3de95c2d WatchSource:0}: Error finding container a43cff83e2e19f182e5b5e1e17e35d28c0cea02c7fff97a3c46f427a3de95c2d: Status 404 returned error can't find the container with id a43cff83e2e19f182e5b5e1e17e35d28c0cea02c7fff97a3c46f427a3de95c2d Oct 04 03:12:56 crc kubenswrapper[4742]: W1004 03:12:56.677040 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62ad7c0a_a108_4c0f_b998_1accb11feb51.slice/crio-eb6ace8f8542d7ffa06eae241f232358bb063fae8ddc7d8d0364f1d480818c14 WatchSource:0}: Error finding container eb6ace8f8542d7ffa06eae241f232358bb063fae8ddc7d8d0364f1d480818c14: Status 404 returned error can't find the container with id eb6ace8f8542d7ffa06eae241f232358bb063fae8ddc7d8d0364f1d480818c14 Oct 04 03:12:56 crc kubenswrapper[4742]: W1004 03:12:56.677326 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8edba022_cd27_43b2_87e3_a0e92562c450.slice/crio-850988efe5b3752e5f766f2cfaeb72e0f4581f0bc86e183604d7b103677f4241 WatchSource:0}: Error finding container 850988efe5b3752e5f766f2cfaeb72e0f4581f0bc86e183604d7b103677f4241: Status 404 returned error can't find the container with id 850988efe5b3752e5f766f2cfaeb72e0f4581f0bc86e183604d7b103677f4241 Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.680003 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-h22jn" event={"ID":"0e5511f2-7587-4638-bd0c-af91c57a0163","Type":"ContainerStarted","Data":"d975d272f7dbbfe0ebee90accbb3ccbf015123ea200150c340b8da79b572569e"} Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.685345 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-gszdm" Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.718665 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6nmn2"] Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.725578 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:12:56 crc kubenswrapper[4742]: E1004 03:12:56.726658 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:12:57.226617356 +0000 UTC m=+143.158954532 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.778112 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qznqm"] Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.790541 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325780-9b22t"] Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.822763 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xg5l9"] Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.828345 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:56 crc kubenswrapper[4742]: E1004 03:12:56.830575 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:12:57.330558922 +0000 UTC m=+143.262896098 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.887062 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-nvt5z"] Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.887120 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-krpzs"] Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.930633 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:12:56 crc kubenswrapper[4742]: E1004 03:12:56.931327 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:12:57.43126195 +0000 UTC m=+143.363599126 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:56 crc kubenswrapper[4742]: I1004 03:12:56.998232 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-8w75v"] Oct 04 03:12:57 crc kubenswrapper[4742]: W1004 03:12:57.017598 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7f11b3c3_6380_4f4d_9d03_622e66e33bc8.slice/crio-f316ca86005256022008feff320c8d58a736775c9530c7176468953d9853697a WatchSource:0}: Error finding container f316ca86005256022008feff320c8d58a736775c9530c7176468953d9853697a: Status 404 returned error can't find the container with id f316ca86005256022008feff320c8d58a736775c9530c7176468953d9853697a Oct 04 03:12:57 crc kubenswrapper[4742]: W1004 03:12:57.028979 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a41af56_8a6d_49ba_8b85_1c546d49c198.slice/crio-2818d69c2bcde7c2a97c921fe134728c7a4bb879085ef5b957f6701cae1e6ed3 WatchSource:0}: Error finding container 2818d69c2bcde7c2a97c921fe134728c7a4bb879085ef5b957f6701cae1e6ed3: Status 404 returned error can't find the container with id 2818d69c2bcde7c2a97c921fe134728c7a4bb879085ef5b957f6701cae1e6ed3 Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.032519 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:57 crc kubenswrapper[4742]: E1004 03:12:57.032904 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:12:57.532885423 +0000 UTC m=+143.465222599 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:57 crc kubenswrapper[4742]: W1004 03:12:57.071101 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod530fb6b6_9c36_4d13_8aa9_903d7080082f.slice/crio-dcf1d53f08519b4bea2b86689fc3b954978b9af7317787f5b92cb0a90f487210 WatchSource:0}: Error finding container dcf1d53f08519b4bea2b86689fc3b954978b9af7317787f5b92cb0a90f487210: Status 404 returned error can't find the container with id dcf1d53f08519b4bea2b86689fc3b954978b9af7317787f5b92cb0a90f487210 Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.134255 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:12:57 crc kubenswrapper[4742]: E1004 03:12:57.134778 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:12:57.634755313 +0000 UTC m=+143.567092489 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.172537 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-fcc5d" podStartSLOduration=119.172514534 podStartE2EDuration="1m59.172514534s" podCreationTimestamp="2025-10-04 03:10:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:12:57.134790654 +0000 UTC m=+143.067127830" watchObservedRunningTime="2025-10-04 03:12:57.172514534 +0000 UTC m=+143.104851720" Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.210548 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-wn8jl"] Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.247753 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-lgfzr"] Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.247290 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:57 crc kubenswrapper[4742]: E1004 03:12:57.260537 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:12:57.760508992 +0000 UTC m=+143.692846168 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.267478 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rtd5w"] Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.272201 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-cms9s"] Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.347394 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t7pnt" podStartSLOduration=119.34737047 podStartE2EDuration="1m59.34737047s" podCreationTimestamp="2025-10-04 03:10:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:12:57.342003286 +0000 UTC m=+143.274340482" watchObservedRunningTime="2025-10-04 03:12:57.34737047 +0000 UTC m=+143.279707656" Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.363565 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:12:57 crc kubenswrapper[4742]: E1004 03:12:57.368307 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:12:57.868241599 +0000 UTC m=+143.800578775 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.402378 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-k77jn"] Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.427525 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8kfz8"] Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.453916 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-p559m"] Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.471723 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:57 crc kubenswrapper[4742]: E1004 03:12:57.472478 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:12:57.972462982 +0000 UTC m=+143.904800158 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.507233 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-6prnz"] Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.513603 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-88xxl" podStartSLOduration=120.513569903 podStartE2EDuration="2m0.513569903s" podCreationTimestamp="2025-10-04 03:10:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:12:57.499385633 +0000 UTC m=+143.431722809" watchObservedRunningTime="2025-10-04 03:12:57.513569903 +0000 UTC m=+143.445907079" Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.573266 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:12:57 crc kubenswrapper[4742]: E1004 03:12:57.573690 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:12:58.073665163 +0000 UTC m=+144.006002339 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.579713 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-gszdm" podStartSLOduration=120.579695885 podStartE2EDuration="2m0.579695885s" podCreationTimestamp="2025-10-04 03:10:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:12:57.577248959 +0000 UTC m=+143.509586135" watchObservedRunningTime="2025-10-04 03:12:57.579695885 +0000 UTC m=+143.512033061" Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.675523 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:57 crc kubenswrapper[4742]: E1004 03:12:57.676443 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:12:58.176430227 +0000 UTC m=+144.108767393 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.737608 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rlbvb" event={"ID":"8edba022-cd27-43b2-87e3-a0e92562c450","Type":"ContainerStarted","Data":"850988efe5b3752e5f766f2cfaeb72e0f4581f0bc86e183604d7b103677f4241"} Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.766940 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tfc7g"] Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.780471 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:12:57 crc kubenswrapper[4742]: E1004 03:12:57.780840 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:12:58.280820114 +0000 UTC m=+144.213157290 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.799566 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-p559m" event={"ID":"bf0e3a33-4470-430f-8ebc-71d52b649684","Type":"ContainerStarted","Data":"9bf062c271833307805b1900c3b5f4e4cafd694ba12173002f73927371e11eac"} Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.802353 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kfp77"] Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.839084 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" event={"ID":"56722804-fa7c-4639-a4d3-5085cd965f8b","Type":"ContainerStarted","Data":"cc9433abc93ed4eadb18b6ea683c8985ffe75711073814e6e91e1da79600208a"} Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.844881 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.853825 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-ktnrj"] Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.859860 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ncmp2" event={"ID":"62ad7c0a-a108-4c0f-b998-1accb11feb51","Type":"ContainerStarted","Data":"eb6ace8f8542d7ffa06eae241f232358bb063fae8ddc7d8d0364f1d480818c14"} Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.866446 4742 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-mspvb container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.9:6443/healthz\": dial tcp 10.217.0.9:6443: connect: connection refused" start-of-body= Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.866514 4742 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" podUID="56722804-fa7c-4639-a4d3-5085cd965f8b" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.9:6443/healthz\": dial tcp 10.217.0.9:6443: connect: connection refused" Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.871841 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-cms9s" event={"ID":"528adbb9-98ba-45d3-ba3a-2fe715e65cf6","Type":"ContainerStarted","Data":"33e2d1c288bfc36eaa9050d576486e348ccbe91e4f04e154e9890e2a9e8275d9"} Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.894112 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" podStartSLOduration=120.894090479 podStartE2EDuration="2m0.894090479s" podCreationTimestamp="2025-10-04 03:10:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:12:57.872923422 +0000 UTC m=+143.805260598" watchObservedRunningTime="2025-10-04 03:12:57.894090479 +0000 UTC m=+143.826427655" Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.897118 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:57 crc kubenswrapper[4742]: E1004 03:12:57.899298 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:12:58.399260628 +0000 UTC m=+144.331597804 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.900241 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-h22jn" event={"ID":"0e5511f2-7587-4638-bd0c-af91c57a0163","Type":"ContainerStarted","Data":"c19ce7f371154b7efa478f08185b50b53092ff7fe11024666fe0c15914d43392"} Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.906945 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-v92dt" event={"ID":"d33e79f4-4871-4e03-ac04-2000ceb3416d","Type":"ContainerStarted","Data":"cb42f04d6bf27a3044735b0f0b83a1f2339e52a4597a1bc58a2acc0d249b3441"} Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.911680 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325780-9b22t" event={"ID":"7f11b3c3-6380-4f4d-9d03-622e66e33bc8","Type":"ContainerStarted","Data":"f316ca86005256022008feff320c8d58a736775c9530c7176468953d9853697a"} Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.942035 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-h22jn" Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.956030 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q" event={"ID":"6d53af45-9fb2-4d36-8c21-aa167c8ed70f","Type":"ContainerStarted","Data":"a43cff83e2e19f182e5b5e1e17e35d28c0cea02c7fff97a3c46f427a3de95c2d"} Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.956115 4742 patch_prober.go:28] interesting pod/downloads-7954f5f757-h22jn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.956170 4742 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-h22jn" podUID="0e5511f2-7587-4638-bd0c-af91c57a0163" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.968167 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5j88f"] Oct 04 03:12:57 crc kubenswrapper[4742]: I1004 03:12:57.997558 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:12:57 crc kubenswrapper[4742]: E1004 03:12:57.998522 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:12:58.498472656 +0000 UTC m=+144.430809832 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:58 crc kubenswrapper[4742]: I1004 03:12:58.050392 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xg5l9" event={"ID":"80a23793-4622-4288-a5c0-de5199e1e767","Type":"ContainerStarted","Data":"15684303d19a3eabf7f3a11a7be7d6d0eb3c653ea9a57e86d59c8909bcd4c422"} Oct 04 03:12:58 crc kubenswrapper[4742]: I1004 03:12:58.067488 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8kfz8" event={"ID":"7336d2fb-f015-4e22-8419-fdefc49f6137","Type":"ContainerStarted","Data":"1f4e789c09116e09cd8ee3f84e954f587b8058f6f86fd6ad745933ebe78e8768"} Oct 04 03:12:58 crc kubenswrapper[4742]: I1004 03:12:58.070991 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-lgfzr" event={"ID":"2556513e-2080-4ed1-99cc-bb1e15a75a8e","Type":"ContainerStarted","Data":"e42a321a4d9ff55243dd8a2ac2e77121232c3057569ae85a0065370ab4dd8e09"} Oct 04 03:12:58 crc kubenswrapper[4742]: I1004 03:12:58.081000 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-krpzs" event={"ID":"530fb6b6-9c36-4d13-8aa9-903d7080082f","Type":"ContainerStarted","Data":"dcf1d53f08519b4bea2b86689fc3b954978b9af7317787f5b92cb0a90f487210"} Oct 04 03:12:58 crc kubenswrapper[4742]: I1004 03:12:58.086070 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rtd5w" event={"ID":"912dc882-2c54-4a47-b26b-6f3a63f9f33f","Type":"ContainerStarted","Data":"941eff9e243b4851c5d4b4f59c2cc2943c227cd9e91d13bfc29bc3230501cadf"} Oct 04 03:12:58 crc kubenswrapper[4742]: W1004 03:12:58.091614 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podadc0635a_397c_48ba_b103_a8d6e62ae535.slice/crio-63442e3a65626ce8a8d37c59c195f7924c7a44cea1a527f22676c2dbf5ac1e58 WatchSource:0}: Error finding container 63442e3a65626ce8a8d37c59c195f7924c7a44cea1a527f22676c2dbf5ac1e58: Status 404 returned error can't find the container with id 63442e3a65626ce8a8d37c59c195f7924c7a44cea1a527f22676c2dbf5ac1e58 Oct 04 03:12:58 crc kubenswrapper[4742]: I1004 03:12:58.095557 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8w75v" event={"ID":"bb08fe29-5abf-43af-a622-f1db8e955e55","Type":"ContainerStarted","Data":"36ef6a3e23cac4bfe91cc3cb8d820153a1f2c4227905522a3eab5c7720128789"} Oct 04 03:12:58 crc kubenswrapper[4742]: I1004 03:12:58.097904 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mgh2f" event={"ID":"c975f343-d9c0-4145-b0b8-bdc11dbd1d66","Type":"ContainerStarted","Data":"6bd7650cd4fee04c0dc80a2790168b158e6507969af24698f12d0d3464b82316"} Oct 04 03:12:58 crc kubenswrapper[4742]: I1004 03:12:58.100959 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:58 crc kubenswrapper[4742]: E1004 03:12:58.101469 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:12:58.601450646 +0000 UTC m=+144.533787822 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:58 crc kubenswrapper[4742]: I1004 03:12:58.101791 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-f5978" event={"ID":"e537cd22-13ae-4f86-a246-aa53e62e8e4d","Type":"ContainerStarted","Data":"8b2a3f30e42f8419c1f38f95e0f1d5b060c4f6eaae0f09f1c77203f5ffb4d41b"} Oct 04 03:12:58 crc kubenswrapper[4742]: I1004 03:12:58.104512 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pkdzv" event={"ID":"da95bbc1-a1c0-428f-9884-cda886a8da3c","Type":"ContainerStarted","Data":"6ef61813e69e4a3ced99489fea9e753a691fb1f6b1fe2de73fae6f5a876b421d"} Oct 04 03:12:58 crc kubenswrapper[4742]: I1004 03:12:58.110115 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-4v795" event={"ID":"77153235-6efa-4db2-a4a3-5153d9ba3568","Type":"ContainerStarted","Data":"6248be10e0be187822195da429c8587aadbd68ec7e2b8c7b17edcf9a01f8463a"} Oct 04 03:12:58 crc kubenswrapper[4742]: I1004 03:12:58.129199 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-2pdbs" event={"ID":"07f1cf4a-506e-469a-a8e5-0adc68a5ab81","Type":"ContainerStarted","Data":"c3cea91521fa1c828a478d157a3534f3d3feaad1cc1532b57a1cd72dbeb5776d"} Oct 04 03:12:58 crc kubenswrapper[4742]: I1004 03:12:58.130523 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-2pdbs" Oct 04 03:12:58 crc kubenswrapper[4742]: I1004 03:12:58.142523 4742 patch_prober.go:28] interesting pod/console-operator-58897d9998-2pdbs container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/readyz\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Oct 04 03:12:58 crc kubenswrapper[4742]: I1004 03:12:58.142595 4742 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-2pdbs" podUID="07f1cf4a-506e-469a-a8e5-0adc68a5ab81" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.21:8443/readyz\": dial tcp 10.217.0.21:8443: connect: connection refused" Oct 04 03:12:58 crc kubenswrapper[4742]: I1004 03:12:58.145377 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6nmn2" event={"ID":"5b2832d6-470e-4f3d-bec4-83012fb6e541","Type":"ContainerStarted","Data":"d33d684d9a4df760c3ad206ed3e9bb6d0bb9694ad4a80854e85e9f7467164f7a"} Oct 04 03:12:58 crc kubenswrapper[4742]: I1004 03:12:58.150488 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wn8jl" event={"ID":"35542dc5-96db-4eca-a2c2-85ebf65ea31f","Type":"ContainerStarted","Data":"974c85402a51e10cdd3ecf77814f8848cd5e8c286f61e040592a7a369e27f601"} Oct 04 03:12:58 crc kubenswrapper[4742]: I1004 03:12:58.162014 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qznqm" event={"ID":"9d4974fb-9d3c-4e2d-ac54-7c3d4b4b3f89","Type":"ContainerStarted","Data":"c2b08c84d7ed18149dc6bd2254e149b1d5b2498dbfcc9c67ea12270b0b5cee23"} Oct 04 03:12:58 crc kubenswrapper[4742]: I1004 03:12:58.180375 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-dmd8c" event={"ID":"7940c6dd-4863-4c92-ad0e-27abb961da4a","Type":"ContainerStarted","Data":"f7d70a6ae4e354e2b217b767ceef38b9c2628aa10d6ef40a57e7d068a28c1887"} Oct 04 03:12:58 crc kubenswrapper[4742]: I1004 03:12:58.187494 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-vtmjj"] Oct 04 03:12:58 crc kubenswrapper[4742]: I1004 03:12:58.197735 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-85fhx" event={"ID":"093665e3-6dc0-4c51-a0f3-71363eef8367","Type":"ContainerStarted","Data":"23803c8a809c98feacb5dbd11bbb49c437ceeaab36cb807ffac7508c4fbfd11f"} Oct 04 03:12:58 crc kubenswrapper[4742]: I1004 03:12:58.202987 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:12:58 crc kubenswrapper[4742]: E1004 03:12:58.204654 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:12:58.70463238 +0000 UTC m=+144.636969556 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:58 crc kubenswrapper[4742]: I1004 03:12:58.242217 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ctkcp" event={"ID":"5ba77307-01a4-43a3-8888-3da1c700ea2e","Type":"ContainerStarted","Data":"049360e0c01238f38b8b2894037d8af7beb23f47451af49453d217ec2f8c1f58"} Oct 04 03:12:58 crc kubenswrapper[4742]: I1004 03:12:58.250751 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-k77jn" event={"ID":"cd220c67-e881-4180-b1c4-4064be0794ca","Type":"ContainerStarted","Data":"d218667585c286664531573754f0c51d9fc17ebd4a1f401f3c2394aff2748dc1"} Oct 04 03:12:58 crc kubenswrapper[4742]: I1004 03:12:58.265059 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-nvt5z" event={"ID":"3a41af56-8a6d-49ba-8b85-1c546d49c198","Type":"ContainerStarted","Data":"2818d69c2bcde7c2a97c921fe134728c7a4bb879085ef5b957f6701cae1e6ed3"} Oct 04 03:12:58 crc kubenswrapper[4742]: I1004 03:12:58.311380 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:58 crc kubenswrapper[4742]: E1004 03:12:58.313954 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:12:58.813934839 +0000 UTC m=+144.746272015 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:58 crc kubenswrapper[4742]: I1004 03:12:58.415003 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:12:58 crc kubenswrapper[4742]: E1004 03:12:58.415586 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:12:58.915564342 +0000 UTC m=+144.847901518 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:58 crc kubenswrapper[4742]: I1004 03:12:58.478398 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-h22jn" podStartSLOduration=121.478379256 podStartE2EDuration="2m1.478379256s" podCreationTimestamp="2025-10-04 03:10:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:12:58.47666712 +0000 UTC m=+144.409004296" watchObservedRunningTime="2025-10-04 03:12:58.478379256 +0000 UTC m=+144.410716422" Oct 04 03:12:58 crc kubenswrapper[4742]: I1004 03:12:58.480112 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-ctkcp" podStartSLOduration=121.480104192 podStartE2EDuration="2m1.480104192s" podCreationTimestamp="2025-10-04 03:10:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:12:58.452572474 +0000 UTC m=+144.384909650" watchObservedRunningTime="2025-10-04 03:12:58.480104192 +0000 UTC m=+144.412441368" Oct 04 03:12:58 crc kubenswrapper[4742]: I1004 03:12:58.516794 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:58 crc kubenswrapper[4742]: E1004 03:12:58.517348 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:12:59.017326869 +0000 UTC m=+144.949664045 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:58 crc kubenswrapper[4742]: I1004 03:12:58.539520 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-2pdbs" podStartSLOduration=121.539489533 podStartE2EDuration="2m1.539489533s" podCreationTimestamp="2025-10-04 03:10:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:12:58.53044219 +0000 UTC m=+144.462779366" watchObservedRunningTime="2025-10-04 03:12:58.539489533 +0000 UTC m=+144.471826699" Oct 04 03:12:58 crc kubenswrapper[4742]: I1004 03:12:58.571133 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-mgh2f" podStartSLOduration=121.57111138 podStartE2EDuration="2m1.57111138s" podCreationTimestamp="2025-10-04 03:10:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:12:58.567832392 +0000 UTC m=+144.500169558" watchObservedRunningTime="2025-10-04 03:12:58.57111138 +0000 UTC m=+144.503448556" Oct 04 03:12:58 crc kubenswrapper[4742]: I1004 03:12:58.618253 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:12:58 crc kubenswrapper[4742]: E1004 03:12:58.618684 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:12:59.118665354 +0000 UTC m=+145.051002530 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:58 crc kubenswrapper[4742]: I1004 03:12:58.722478 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:58 crc kubenswrapper[4742]: E1004 03:12:58.722775 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:12:59.222764043 +0000 UTC m=+145.155101219 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:58 crc kubenswrapper[4742]: I1004 03:12:58.822923 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:12:58 crc kubenswrapper[4742]: E1004 03:12:58.823387 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:12:59.323362888 +0000 UTC m=+145.255700074 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:58 crc kubenswrapper[4742]: I1004 03:12:58.924236 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:58 crc kubenswrapper[4742]: E1004 03:12:58.924655 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:12:59.424637872 +0000 UTC m=+145.356975048 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.030716 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:12:59 crc kubenswrapper[4742]: E1004 03:12:59.031248 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:12:59.531224928 +0000 UTC m=+145.463562104 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.132851 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:59 crc kubenswrapper[4742]: E1004 03:12:59.133298 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:12:59.633264342 +0000 UTC m=+145.565601508 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.252807 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:12:59 crc kubenswrapper[4742]: E1004 03:12:59.253201 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:12:59.753182096 +0000 UTC m=+145.685519272 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.306679 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-v92dt" event={"ID":"d33e79f4-4871-4e03-ac04-2000ceb3416d","Type":"ContainerStarted","Data":"be81bedd999a94e83a7b7b2438621408c1edd618d6ff35df25acd6688bb2261a"} Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.322618 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6nmn2" event={"ID":"5b2832d6-470e-4f3d-bec4-83012fb6e541","Type":"ContainerStarted","Data":"4630ec576bcacb3ccd72b650283195c79c90e185fd4704aa8f413f3379addd9d"} Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.331891 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rlbvb" event={"ID":"8edba022-cd27-43b2-87e3-a0e92562c450","Type":"ContainerStarted","Data":"765a3fb433b61962a95804d97c036d7773bab89f592ff024d1a716307557781a"} Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.355105 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:59 crc kubenswrapper[4742]: E1004 03:12:59.357749 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:12:59.857726397 +0000 UTC m=+145.790063573 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.360458 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-fvr9s" event={"ID":"09503673-286e-48af-b18b-37bff87f8019","Type":"ContainerStarted","Data":"5b31bc0a0d668d4e1e0bde3bc04c02a05b3758d62d618ea8cf6c77805f7eaa9a"} Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.369067 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qznqm" event={"ID":"9d4974fb-9d3c-4e2d-ac54-7c3d4b4b3f89","Type":"ContainerStarted","Data":"4ad52d3369219c69476baa30974e5602927c20c0bd50b26e9581ed9cd043995d"} Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.372422 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-v92dt" podStartSLOduration=122.37240524 podStartE2EDuration="2m2.37240524s" podCreationTimestamp="2025-10-04 03:10:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:12:59.339074437 +0000 UTC m=+145.271411613" watchObservedRunningTime="2025-10-04 03:12:59.37240524 +0000 UTC m=+145.304742416" Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.373494 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6nmn2" podStartSLOduration=122.373485409 podStartE2EDuration="2m2.373485409s" podCreationTimestamp="2025-10-04 03:10:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:12:59.370732535 +0000 UTC m=+145.303069711" watchObservedRunningTime="2025-10-04 03:12:59.373485409 +0000 UTC m=+145.305822585" Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.377416 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8kfz8" event={"ID":"7336d2fb-f015-4e22-8419-fdefc49f6137","Type":"ContainerStarted","Data":"fb6788978bb8a66df98e695a2af1722ee2ae66a351671bd27f8782e74c84db49"} Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.378179 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8kfz8" Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.389563 4742 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-8kfz8 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" start-of-body= Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.389639 4742 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8kfz8" podUID="7336d2fb-f015-4e22-8419-fdefc49f6137" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.405524 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-p559m" event={"ID":"bf0e3a33-4470-430f-8ebc-71d52b649684","Type":"ContainerStarted","Data":"a17a6547faf450968d1c8b2e712d4ff55acb427c201f161e3539d0885a639f18"} Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.429746 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rtd5w" event={"ID":"912dc882-2c54-4a47-b26b-6f3a63f9f33f","Type":"ContainerStarted","Data":"21909438800c70298ffc8c7eab5972974662e95ada9c15f037953acb5a1ed0fa"} Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.432473 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rtd5w" Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.438543 4742 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-rtd5w container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:5443/healthz\": dial tcp 10.217.0.32:5443: connect: connection refused" start-of-body= Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.438622 4742 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rtd5w" podUID="912dc882-2c54-4a47-b26b-6f3a63f9f33f" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.32:5443/healthz\": dial tcp 10.217.0.32:5443: connect: connection refused" Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.461306 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:12:59 crc kubenswrapper[4742]: E1004 03:12:59.462849 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:12:59.962826533 +0000 UTC m=+145.895163709 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.495400 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rlbvb" podStartSLOduration=121.495377745 podStartE2EDuration="2m1.495377745s" podCreationTimestamp="2025-10-04 03:10:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:12:59.406085343 +0000 UTC m=+145.338422519" watchObservedRunningTime="2025-10-04 03:12:59.495377745 +0000 UTC m=+145.427714911" Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.556871 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-krpzs" event={"ID":"530fb6b6-9c36-4d13-8aa9-903d7080082f","Type":"ContainerStarted","Data":"4c616fd94b3171e07727bc019fa23d1d225dc213845c0496f10a9764bcd79b39"} Oct 04 03:12:59 crc kubenswrapper[4742]: E1004 03:12:59.567110 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:13:00.066371498 +0000 UTC m=+145.998708674 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.565911 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.567157 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-krpzs" event={"ID":"530fb6b6-9c36-4d13-8aa9-903d7080082f","Type":"ContainerStarted","Data":"40a29011dba3850ac7b7ecd3b2548278ff60f8999757889be36e37eb8087faa7"} Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.596586 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-85fhx" event={"ID":"093665e3-6dc0-4c51-a0f3-71363eef8367","Type":"ContainerStarted","Data":"81468d8fc95b7db0fb578cec455bbbbe5959062bbfae4e36bc67317ea1d80bd4"} Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.599951 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mfp6b" event={"ID":"e3b81c30-6a85-430c-a915-6322d20953ac","Type":"ContainerStarted","Data":"c32136db50fc02ba8c559a6816b1a258a1a4136c87ac9735ac79d69166911f0a"} Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.600009 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mfp6b" event={"ID":"e3b81c30-6a85-430c-a915-6322d20953ac","Type":"ContainerStarted","Data":"4a3014b11f26e8f4a8c86a785247b7fe1d6146519207989c5d93c9939d11a4ed"} Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.614747 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pkdzv" event={"ID":"da95bbc1-a1c0-428f-9884-cda886a8da3c","Type":"ContainerStarted","Data":"388f6c303f97d464bd9047c5977d1c0f003da3a83d669308fcae9bbec91ba419"} Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.633253 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8kfz8" podStartSLOduration=121.633232609 podStartE2EDuration="2m1.633232609s" podCreationTimestamp="2025-10-04 03:10:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:12:59.538208003 +0000 UTC m=+145.470545179" watchObservedRunningTime="2025-10-04 03:12:59.633232609 +0000 UTC m=+145.565569785" Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.664507 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-nvt5z" event={"ID":"3a41af56-8a6d-49ba-8b85-1c546d49c198","Type":"ContainerStarted","Data":"7207f6f1cb30217fe6721e55b1fb3f637bdb281cb7aa417da9b9904e696910d4"} Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.669366 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:12:59 crc kubenswrapper[4742]: E1004 03:12:59.671926 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:13:00.171904485 +0000 UTC m=+146.104241661 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.709405 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-fvr9s" podStartSLOduration=122.70938317 podStartE2EDuration="2m2.70938317s" podCreationTimestamp="2025-10-04 03:10:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:12:59.651504099 +0000 UTC m=+145.583841275" watchObservedRunningTime="2025-10-04 03:12:59.70938317 +0000 UTC m=+145.641720366" Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.717436 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pkdzv" podStartSLOduration=122.717399415 podStartE2EDuration="2m2.717399415s" podCreationTimestamp="2025-10-04 03:10:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:12:59.708230379 +0000 UTC m=+145.640567565" watchObservedRunningTime="2025-10-04 03:12:59.717399415 +0000 UTC m=+145.649736601" Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.727589 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ncmp2" event={"ID":"62ad7c0a-a108-4c0f-b998-1accb11feb51","Type":"ContainerStarted","Data":"3651de55447c17eb0903858d4a0f1df0455657a85a2b4169d4a4bd0b9d022973"} Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.770675 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6prnz" event={"ID":"4ba39c3a-d31f-4f1d-807d-c9fcf6defa34","Type":"ContainerStarted","Data":"2c1e61105532c8aaad284748a5e54b701dbb48e3ce8b84f8fb490e01bd2ba834"} Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.770729 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6prnz" event={"ID":"4ba39c3a-d31f-4f1d-807d-c9fcf6defa34","Type":"ContainerStarted","Data":"8fe7822612143fc55baf97b9ac12a44fb3127b7f7dda4df261784bc8c7932de4"} Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.771439 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:59 crc kubenswrapper[4742]: E1004 03:12:59.772580 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:13:00.272558733 +0000 UTC m=+146.204895909 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.776532 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mfp6b" podStartSLOduration=122.776505358 podStartE2EDuration="2m2.776505358s" podCreationTimestamp="2025-10-04 03:10:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:12:59.761191528 +0000 UTC m=+145.693528704" watchObservedRunningTime="2025-10-04 03:12:59.776505358 +0000 UTC m=+145.708842534" Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.803619 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-4v795" event={"ID":"77153235-6efa-4db2-a4a3-5153d9ba3568","Type":"ContainerStarted","Data":"2ef3d5edd3fae17309687975c57a63b5533b2b9425b0377313980473cf9de79c"} Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.806732 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-dmd8c" event={"ID":"7940c6dd-4863-4c92-ad0e-27abb961da4a","Type":"ContainerStarted","Data":"3016f70e32e30f5a4ee5e24a0c5feff127ce514998fd278009faeba31b4f0a32"} Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.808590 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-lgfzr" event={"ID":"2556513e-2080-4ed1-99cc-bb1e15a75a8e","Type":"ContainerStarted","Data":"c1d9f7143072fa6b1ccc510506e5e04cb743a74fc48bb61c1433adecd1f7e10c"} Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.816896 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-vtmjj" event={"ID":"c151b75e-6548-4d28-851e-d8f4eac68183","Type":"ContainerStarted","Data":"a531f58dfa633abfe30fcf54dbf078efd80dcd76ac81fbf2e20117a26f20154c"} Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.851417 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5j88f" event={"ID":"adc0635a-397c-48ba-b103-a8d6e62ae535","Type":"ContainerStarted","Data":"f2849a140d54f6af13b5d8e30b4657c11e2cefda44367f7da94c7c34e7b90866"} Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.851467 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5j88f" event={"ID":"adc0635a-397c-48ba-b103-a8d6e62ae535","Type":"ContainerStarted","Data":"63442e3a65626ce8a8d37c59c195f7924c7a44cea1a527f22676c2dbf5ac1e58"} Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.852178 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5j88f" Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.875164 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:12:59 crc kubenswrapper[4742]: E1004 03:12:59.876511 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:13:00.376489567 +0000 UTC m=+146.308826743 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.883483 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nsspd" event={"ID":"8725e189-8b15-4c13-ae9b-f1230476e572","Type":"ContainerStarted","Data":"19ace0ce93872eb807fc6008fe6b5b87eb700872ed54d969099afc45cba5fa1c"} Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.884333 4742 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-5j88f container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" start-of-body= Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.884378 4742 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5j88f" podUID="adc0635a-397c-48ba-b103-a8d6e62ae535" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.897746 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-85fhx" podStartSLOduration=122.897726736 podStartE2EDuration="2m2.897726736s" podCreationTimestamp="2025-10-04 03:10:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:12:59.862234615 +0000 UTC m=+145.794571791" watchObservedRunningTime="2025-10-04 03:12:59.897726736 +0000 UTC m=+145.830063912" Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.897929 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-krpzs" podStartSLOduration=121.897925182 podStartE2EDuration="2m1.897925182s" podCreationTimestamp="2025-10-04 03:10:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:12:59.89711721 +0000 UTC m=+145.829454386" watchObservedRunningTime="2025-10-04 03:12:59.897925182 +0000 UTC m=+145.830262358" Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.901084 4742 generic.go:334] "Generic (PLEG): container finished" podID="6d53af45-9fb2-4d36-8c21-aa167c8ed70f" containerID="90615d5acc5ed66a92528514c38b0bf7a523af266bcfcdef9b0e0c97646cdedb" exitCode=0 Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.901184 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q" event={"ID":"6d53af45-9fb2-4d36-8c21-aa167c8ed70f","Type":"ContainerDied","Data":"90615d5acc5ed66a92528514c38b0bf7a523af266bcfcdef9b0e0c97646cdedb"} Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.930726 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-f5978" event={"ID":"e537cd22-13ae-4f86-a246-aa53e62e8e4d","Type":"ContainerStarted","Data":"c91811362103fb0ef5589fe91cc8db9321ba678fff3d6104e195415d5d9eda2d"} Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.952666 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tfc7g" event={"ID":"a8b41c08-4e89-4f08-8d45-256c306fbea0","Type":"ContainerStarted","Data":"df9ce9e210f6d14e489c7061ca2d98d8972dc36b37ba107b79b69f93f09038ec"} Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.954008 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-tfc7g" Oct 04 03:12:59 crc kubenswrapper[4742]: I1004 03:12:59.980045 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:12:59 crc kubenswrapper[4742]: E1004 03:12:59.982559 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:13:00.482542799 +0000 UTC m=+146.414879975 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.001912 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xg5l9" event={"ID":"80a23793-4622-4288-a5c0-de5199e1e767","Type":"ContainerStarted","Data":"b6e71c15d1b3c0221be900f22ee943b24c1ea0014ce094c28c814d232080f735"} Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.001983 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xg5l9" event={"ID":"80a23793-4622-4288-a5c0-de5199e1e767","Type":"ContainerStarted","Data":"dd4613cbf6ea1393adb32a17976a17331ded6b061e963030979c5ebaf84fe02b"} Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.002816 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xg5l9" Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.002920 4742 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-tfc7g container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.002955 4742 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-tfc7g" podUID="a8b41c08-4e89-4f08-8d45-256c306fbea0" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.039532 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rtd5w" podStartSLOduration=122.039512766 podStartE2EDuration="2m2.039512766s" podCreationTimestamp="2025-10-04 03:10:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:12:59.994622723 +0000 UTC m=+145.926959909" watchObservedRunningTime="2025-10-04 03:13:00.039512766 +0000 UTC m=+145.971849942" Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.049403 4742 generic.go:334] "Generic (PLEG): container finished" podID="bb08fe29-5abf-43af-a622-f1db8e955e55" containerID="3a9ae1b1c0c340f8f19aa230b4cabca32f4bea1db7cdd7fc8c9b054448958889" exitCode=0 Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.049513 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8w75v" event={"ID":"bb08fe29-5abf-43af-a622-f1db8e955e55","Type":"ContainerDied","Data":"3a9ae1b1c0c340f8f19aa230b4cabca32f4bea1db7cdd7fc8c9b054448958889"} Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.082580 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.082966 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-f5978" podStartSLOduration=123.082939959 podStartE2EDuration="2m3.082939959s" podCreationTimestamp="2025-10-04 03:10:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:13:00.040950524 +0000 UTC m=+145.973287720" watchObservedRunningTime="2025-10-04 03:13:00.082939959 +0000 UTC m=+146.015277145" Oct 04 03:13:00 crc kubenswrapper[4742]: E1004 03:13:00.084381 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:13:00.584328146 +0000 UTC m=+146.516665322 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.094573 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325780-9b22t" event={"ID":"7f11b3c3-6380-4f4d-9d03-622e66e33bc8","Type":"ContainerStarted","Data":"5da555775d193aac2a2c45269fd27c648394620f29687dc3df2a5b6c00baa45e"} Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.107163 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-ktnrj" event={"ID":"b16d0ad2-0b45-45fe-bc2c-215697447010","Type":"ContainerStarted","Data":"955b3e42a3b2c435252ac33f0091871b7bec60d2366bca2e8627eb9e1c4ee0c5"} Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.128911 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kfp77" event={"ID":"5057f820-3d60-43e3-9cb8-ceac9650621d","Type":"ContainerStarted","Data":"797f3797a50fd45ce4a4ca698224f6e1fe112bdf789a75f6be2566f37640232f"} Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.132291 4742 patch_prober.go:28] interesting pod/downloads-7954f5f757-h22jn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.132335 4742 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-h22jn" podUID="0e5511f2-7587-4638-bd0c-af91c57a0163" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.140394 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-tfc7g" podStartSLOduration=122.140363858 podStartE2EDuration="2m2.140363858s" podCreationTimestamp="2025-10-04 03:10:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:13:00.135612851 +0000 UTC m=+146.067950037" watchObservedRunningTime="2025-10-04 03:13:00.140363858 +0000 UTC m=+146.072701034" Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.146207 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-85fhx" Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.158497 4742 patch_prober.go:28] interesting pod/router-default-5444994796-85fhx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:13:00 crc kubenswrapper[4742]: [-]has-synced failed: reason withheld Oct 04 03:13:00 crc kubenswrapper[4742]: [+]process-running ok Oct 04 03:13:00 crc kubenswrapper[4742]: healthz check failed Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.158565 4742 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85fhx" podUID="093665e3-6dc0-4c51-a0f3-71363eef8367" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.163992 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-dmd8c" podStartSLOduration=122.163972001 podStartE2EDuration="2m2.163972001s" podCreationTimestamp="2025-10-04 03:10:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:13:00.163797956 +0000 UTC m=+146.096135132" watchObservedRunningTime="2025-10-04 03:13:00.163972001 +0000 UTC m=+146.096309177" Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.184412 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:13:00 crc kubenswrapper[4742]: E1004 03:13:00.187036 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:13:00.687016968 +0000 UTC m=+146.619354144 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.284558 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-4v795" podStartSLOduration=7.284538991 podStartE2EDuration="7.284538991s" podCreationTimestamp="2025-10-04 03:12:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:13:00.223139146 +0000 UTC m=+146.155476352" watchObservedRunningTime="2025-10-04 03:13:00.284538991 +0000 UTC m=+146.216876167" Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.289811 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:13:00 crc kubenswrapper[4742]: E1004 03:13:00.310649 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:13:00.790197473 +0000 UTC m=+146.722534649 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.346626 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-lgfzr" podStartSLOduration=7.346597064 podStartE2EDuration="7.346597064s" podCreationTimestamp="2025-10-04 03:12:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:13:00.286497504 +0000 UTC m=+146.218834680" watchObservedRunningTime="2025-10-04 03:13:00.346597064 +0000 UTC m=+146.278934250" Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.391393 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:13:00 crc kubenswrapper[4742]: E1004 03:13:00.391810 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:13:00.891796225 +0000 UTC m=+146.824133401 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.412171 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nsspd" podStartSLOduration=122.412146641 podStartE2EDuration="2m2.412146641s" podCreationTimestamp="2025-10-04 03:10:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:13:00.410813595 +0000 UTC m=+146.343150791" watchObservedRunningTime="2025-10-04 03:13:00.412146641 +0000 UTC m=+146.344483817" Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.413087 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ncmp2" podStartSLOduration=122.413079356 podStartE2EDuration="2m2.413079356s" podCreationTimestamp="2025-10-04 03:10:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:13:00.346581164 +0000 UTC m=+146.278918340" watchObservedRunningTime="2025-10-04 03:13:00.413079356 +0000 UTC m=+146.345416542" Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.430716 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.495877 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:13:00 crc kubenswrapper[4742]: E1004 03:13:00.496524 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:13:00.996498991 +0000 UTC m=+146.928836177 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.537332 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5j88f" podStartSLOduration=122.537311484 podStartE2EDuration="2m2.537311484s" podCreationTimestamp="2025-10-04 03:10:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:13:00.470671979 +0000 UTC m=+146.403009145" watchObservedRunningTime="2025-10-04 03:13:00.537311484 +0000 UTC m=+146.469648660" Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.554129 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-vtmjj" podStartSLOduration=122.554102654 podStartE2EDuration="2m2.554102654s" podCreationTimestamp="2025-10-04 03:10:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:13:00.53602158 +0000 UTC m=+146.468358756" watchObservedRunningTime="2025-10-04 03:13:00.554102654 +0000 UTC m=+146.486439830" Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.575992 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kfp77" podStartSLOduration=123.57596207 podStartE2EDuration="2m3.57596207s" podCreationTimestamp="2025-10-04 03:10:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:13:00.567986596 +0000 UTC m=+146.500323782" watchObservedRunningTime="2025-10-04 03:13:00.57596207 +0000 UTC m=+146.508299246" Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.598345 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:13:00 crc kubenswrapper[4742]: E1004 03:13:00.598702 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:13:01.098687609 +0000 UTC m=+147.031024785 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.617303 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-ktnrj" podStartSLOduration=122.617250386 podStartE2EDuration="2m2.617250386s" podCreationTimestamp="2025-10-04 03:10:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:13:00.615125319 +0000 UTC m=+146.547462495" watchObservedRunningTime="2025-10-04 03:13:00.617250386 +0000 UTC m=+146.549587562" Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.701893 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:13:00 crc kubenswrapper[4742]: E1004 03:13:00.702557 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:13:01.202524331 +0000 UTC m=+147.134861507 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.729766 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xg5l9" podStartSLOduration=122.729743981 podStartE2EDuration="2m2.729743981s" podCreationTimestamp="2025-10-04 03:10:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:13:00.727746807 +0000 UTC m=+146.660083973" watchObservedRunningTime="2025-10-04 03:13:00.729743981 +0000 UTC m=+146.662081157" Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.759080 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-2pdbs" Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.804735 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:13:00 crc kubenswrapper[4742]: E1004 03:13:00.805221 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:13:01.305204413 +0000 UTC m=+147.237541599 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.889669 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29325780-9b22t" podStartSLOduration=123.889644735 podStartE2EDuration="2m3.889644735s" podCreationTimestamp="2025-10-04 03:10:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:13:00.816443974 +0000 UTC m=+146.748781150" watchObservedRunningTime="2025-10-04 03:13:00.889644735 +0000 UTC m=+146.821981911" Oct 04 03:13:00 crc kubenswrapper[4742]: I1004 03:13:00.908390 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:13:00 crc kubenswrapper[4742]: E1004 03:13:00.909120 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:13:01.409096346 +0000 UTC m=+147.341433522 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.010503 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:13:01 crc kubenswrapper[4742]: E1004 03:13:01.011097 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:13:01.511079559 +0000 UTC m=+147.443416735 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.111333 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:13:01 crc kubenswrapper[4742]: E1004 03:13:01.111563 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:13:01.611528401 +0000 UTC m=+147.543865577 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.111680 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:13:01 crc kubenswrapper[4742]: E1004 03:13:01.112105 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:13:01.612093826 +0000 UTC m=+147.544431172 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.139448 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-nvt5z" event={"ID":"3a41af56-8a6d-49ba-8b85-1c546d49c198","Type":"ContainerStarted","Data":"6c5256af85f80131af1b4a24fe6e6146f99dff00de742f2f84640013f3766f2f"} Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.139603 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-nvt5z" Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.141113 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wn8jl" event={"ID":"35542dc5-96db-4eca-a2c2-85ebf65ea31f","Type":"ContainerStarted","Data":"16fcffd03d2e2b24b9cc820c93dbfaa26cbad9d12b17ec687ce01f0c481e3641"} Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.147388 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tfc7g" event={"ID":"a8b41c08-4e89-4f08-8d45-256c306fbea0","Type":"ContainerStarted","Data":"9f9c46a4ab5b18d70d78e1592b5d352927e2103c6d485583eb027a29fe3d3fa5"} Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.148257 4742 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-tfc7g container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.148345 4742 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-tfc7g" podUID="a8b41c08-4e89-4f08-8d45-256c306fbea0" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.155288 4742 patch_prober.go:28] interesting pod/router-default-5444994796-85fhx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:13:01 crc kubenswrapper[4742]: [-]has-synced failed: reason withheld Oct 04 03:13:01 crc kubenswrapper[4742]: [+]process-running ok Oct 04 03:13:01 crc kubenswrapper[4742]: healthz check failed Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.155387 4742 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85fhx" podUID="093665e3-6dc0-4c51-a0f3-71363eef8367" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.167592 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-cms9s" event={"ID":"528adbb9-98ba-45d3-ba3a-2fe715e65cf6","Type":"ContainerStarted","Data":"42853e9100d12bf408e439d84530b565c540086a13c186de2b775a8bcff1b034"} Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.167644 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-cms9s" event={"ID":"528adbb9-98ba-45d3-ba3a-2fe715e65cf6","Type":"ContainerStarted","Data":"a45cdbb825a33faca72e03a124e3aa2bf16db92ba8d6f5d916dd4a5cb21a8fae"} Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.181936 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kfp77" event={"ID":"5057f820-3d60-43e3-9cb8-ceac9650621d","Type":"ContainerStarted","Data":"64d11a93368a3cee310793f474f96d18dbbd3e9e4d2139441121a83fce545a65"} Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.189295 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-vtmjj" event={"ID":"c151b75e-6548-4d28-851e-d8f4eac68183","Type":"ContainerStarted","Data":"2f403f61d42348667c537614c1c704f6e0d922d74d14c5595c1a4589213e43e5"} Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.190350 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-nvt5z" podStartSLOduration=8.190330752 podStartE2EDuration="8.190330752s" podCreationTimestamp="2025-10-04 03:12:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:13:01.188714849 +0000 UTC m=+147.121052025" watchObservedRunningTime="2025-10-04 03:13:01.190330752 +0000 UTC m=+147.122667928" Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.197247 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-p559m" event={"ID":"bf0e3a33-4470-430f-8ebc-71d52b649684","Type":"ContainerStarted","Data":"038d8fd5b87efcbd1bd06ba41c3078858becaed4cb1897f0a82734cf7b358689"} Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.205913 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8w75v" event={"ID":"bb08fe29-5abf-43af-a622-f1db8e955e55","Type":"ContainerStarted","Data":"5c09bc779ba86a2c8c4e707ff20296c5aeb396826756bfbc2755c7a00cab37f5"} Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.206220 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8w75v" Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.213652 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:13:01 crc kubenswrapper[4742]: E1004 03:13:01.214041 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:13:01.713995236 +0000 UTC m=+147.646332412 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.214511 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:13:01 crc kubenswrapper[4742]: E1004 03:13:01.216319 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:13:01.716295108 +0000 UTC m=+147.648632434 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.231825 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q" event={"ID":"6d53af45-9fb2-4d36-8c21-aa167c8ed70f","Type":"ContainerStarted","Data":"1c6158ebce5f7baee2310ea0e84ceb7a701f7e65abe797023af30abd6bbeffb5"} Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.245657 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-ktnrj" event={"ID":"b16d0ad2-0b45-45fe-bc2c-215697447010","Type":"ContainerStarted","Data":"094dbeb077c4c3b1c109a373feb4a3045b67b65057f16c6f95db5354e1e05091"} Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.288229 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-k77jn" event={"ID":"cd220c67-e881-4180-b1c4-4064be0794ca","Type":"ContainerStarted","Data":"a6fcc50bfb18d6cbfc172df0250f98933129df5af7d6fc1edb69cb06dbbfc4c6"} Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.288310 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-k77jn" event={"ID":"cd220c67-e881-4180-b1c4-4064be0794ca","Type":"ContainerStarted","Data":"e21a19af81cbf55be52e0dd6c218e4eaadcc9eff2a223df28f3dfe20a5cf5060"} Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.298836 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-cms9s" podStartSLOduration=124.298791738 podStartE2EDuration="2m4.298791738s" podCreationTimestamp="2025-10-04 03:10:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:13:01.239969312 +0000 UTC m=+147.172306478" watchObservedRunningTime="2025-10-04 03:13:01.298791738 +0000 UTC m=+147.231128914" Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.305126 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6prnz" event={"ID":"4ba39c3a-d31f-4f1d-807d-c9fcf6defa34","Type":"ContainerStarted","Data":"c3f8f6c41b76d444f245f97230be2d4d0cead54c740e9ccaa62cd9711a5c46b1"} Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.318441 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:13:01 crc kubenswrapper[4742]: E1004 03:13:01.323094 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:13:01.823060949 +0000 UTC m=+147.755398125 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.344653 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qznqm" event={"ID":"9d4974fb-9d3c-4e2d-ac54-7c3d4b4b3f89","Type":"ContainerStarted","Data":"a2938a6bfe61bee9979a7b4ec229224fddd104d90c92147baafdc42dc7c670c4"} Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.356920 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8kfz8" Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.358402 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5j88f" Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.379400 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8w75v" podStartSLOduration=124.379378598 podStartE2EDuration="2m4.379378598s" podCreationTimestamp="2025-10-04 03:10:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:13:01.37870268 +0000 UTC m=+147.311039876" watchObservedRunningTime="2025-10-04 03:13:01.379378598 +0000 UTC m=+147.311715784" Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.384865 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-p559m" podStartSLOduration=123.384823594 podStartE2EDuration="2m3.384823594s" podCreationTimestamp="2025-10-04 03:10:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:13:01.287681411 +0000 UTC m=+147.220018587" watchObservedRunningTime="2025-10-04 03:13:01.384823594 +0000 UTC m=+147.317160780" Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.422936 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:13:01 crc kubenswrapper[4742]: E1004 03:13:01.424232 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:13:01.924210139 +0000 UTC m=+147.856547315 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.424566 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q" podStartSLOduration=123.424546228 podStartE2EDuration="2m3.424546228s" podCreationTimestamp="2025-10-04 03:10:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:13:01.421829855 +0000 UTC m=+147.354167031" watchObservedRunningTime="2025-10-04 03:13:01.424546228 +0000 UTC m=+147.356883404" Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.464820 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6prnz" podStartSLOduration=123.464795467 podStartE2EDuration="2m3.464795467s" podCreationTimestamp="2025-10-04 03:10:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:13:01.464096428 +0000 UTC m=+147.396433614" watchObservedRunningTime="2025-10-04 03:13:01.464795467 +0000 UTC m=+147.397132653" Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.524112 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:13:01 crc kubenswrapper[4742]: E1004 03:13:01.524399 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:13:02.024328132 +0000 UTC m=+147.956665298 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.524799 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:13:01 crc kubenswrapper[4742]: E1004 03:13:01.533028 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:13:02.033003184 +0000 UTC m=+147.965340360 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.546035 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-k77jn" podStartSLOduration=123.546014563 podStartE2EDuration="2m3.546014563s" podCreationTimestamp="2025-10-04 03:10:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:13:01.505863897 +0000 UTC m=+147.438201093" watchObservedRunningTime="2025-10-04 03:13:01.546014563 +0000 UTC m=+147.478351749" Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.583975 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qznqm" podStartSLOduration=124.583947389 podStartE2EDuration="2m4.583947389s" podCreationTimestamp="2025-10-04 03:10:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:13:01.5813527 +0000 UTC m=+147.513689886" watchObservedRunningTime="2025-10-04 03:13:01.583947389 +0000 UTC m=+147.516284565" Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.639940 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:13:01 crc kubenswrapper[4742]: E1004 03:13:01.640502 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:13:02.140474774 +0000 UTC m=+148.072811950 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.743055 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:13:01 crc kubenswrapper[4742]: E1004 03:13:01.744063 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:13:02.244046649 +0000 UTC m=+148.176383825 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.847096 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:13:01 crc kubenswrapper[4742]: E1004 03:13:01.847973 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:13:02.347927953 +0000 UTC m=+148.280265139 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:01 crc kubenswrapper[4742]: I1004 03:13:01.951401 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:13:01 crc kubenswrapper[4742]: E1004 03:13:01.951972 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:13:02.45195072 +0000 UTC m=+148.384287896 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.053523 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:13:02 crc kubenswrapper[4742]: E1004 03:13:02.054319 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:13:02.554292002 +0000 UTC m=+148.486629178 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.151916 4742 patch_prober.go:28] interesting pod/router-default-5444994796-85fhx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:13:02 crc kubenswrapper[4742]: [-]has-synced failed: reason withheld Oct 04 03:13:02 crc kubenswrapper[4742]: [+]process-running ok Oct 04 03:13:02 crc kubenswrapper[4742]: healthz check failed Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.151983 4742 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85fhx" podUID="093665e3-6dc0-4c51-a0f3-71363eef8367" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.156845 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:13:02 crc kubenswrapper[4742]: E1004 03:13:02.157414 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:13:02.657395995 +0000 UTC m=+148.589733171 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.199481 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rtd5w" Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.262867 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:13:02 crc kubenswrapper[4742]: E1004 03:13:02.263176 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:13:02.763134067 +0000 UTC m=+148.695471253 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.263325 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:13:02 crc kubenswrapper[4742]: E1004 03:13:02.264131 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:13:02.764121004 +0000 UTC m=+148.696458190 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.362916 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wn8jl" event={"ID":"35542dc5-96db-4eca-a2c2-85ebf65ea31f","Type":"ContainerStarted","Data":"5aa673d84d32af76037487b8505db84992a0a1d8f3d37f61813fdef97afb53d7"} Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.363292 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wn8jl" event={"ID":"35542dc5-96db-4eca-a2c2-85ebf65ea31f","Type":"ContainerStarted","Data":"c8a0f5e5a51f79ec52565e2f837c8587154f957b67f2210fd62b618ab5f88698"} Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.365113 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:13:02 crc kubenswrapper[4742]: E1004 03:13:02.365823 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:13:02.865805399 +0000 UTC m=+148.798142575 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.366376 4742 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-tfc7g container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.366417 4742 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-tfc7g" podUID="a8b41c08-4e89-4f08-8d45-256c306fbea0" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.466743 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:13:02 crc kubenswrapper[4742]: E1004 03:13:02.469229 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:13:02.969214119 +0000 UTC m=+148.901551295 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.568570 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:13:02 crc kubenswrapper[4742]: E1004 03:13:02.568880 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:13:03.068838769 +0000 UTC m=+149.001175945 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.569299 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:13:02 crc kubenswrapper[4742]: E1004 03:13:02.569690 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:13:03.069674561 +0000 UTC m=+149.002011737 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.670694 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:13:02 crc kubenswrapper[4742]: E1004 03:13:02.670881 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:13:03.170845352 +0000 UTC m=+149.103182528 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.671499 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.671572 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:13:02 crc kubenswrapper[4742]: E1004 03:13:02.672054 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:13:03.172029324 +0000 UTC m=+149.104366510 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.679053 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.773551 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:13:02 crc kubenswrapper[4742]: E1004 03:13:02.773726 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:13:03.273692148 +0000 UTC m=+149.206029324 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.773944 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.774045 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.774127 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.774198 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:13:02 crc kubenswrapper[4742]: E1004 03:13:02.774369 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:13:03.274360796 +0000 UTC m=+149.206697972 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.781616 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.786591 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.787182 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.847109 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-trfnh"] Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.848592 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-trfnh" Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.851927 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.875233 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:13:02 crc kubenswrapper[4742]: E1004 03:13:02.875473 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:13:03.375422984 +0000 UTC m=+149.307760160 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.875540 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zc4gb\" (UniqueName: \"kubernetes.io/projected/2788ab92-5c72-4f27-888e-ef695e4cf6cb-kube-api-access-zc4gb\") pod \"community-operators-trfnh\" (UID: \"2788ab92-5c72-4f27-888e-ef695e4cf6cb\") " pod="openshift-marketplace/community-operators-trfnh" Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.875594 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.875625 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2788ab92-5c72-4f27-888e-ef695e4cf6cb-catalog-content\") pod \"community-operators-trfnh\" (UID: \"2788ab92-5c72-4f27-888e-ef695e4cf6cb\") " pod="openshift-marketplace/community-operators-trfnh" Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.875650 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2788ab92-5c72-4f27-888e-ef695e4cf6cb-utilities\") pod \"community-operators-trfnh\" (UID: \"2788ab92-5c72-4f27-888e-ef695e4cf6cb\") " pod="openshift-marketplace/community-operators-trfnh" Oct 04 03:13:02 crc kubenswrapper[4742]: E1004 03:13:02.876321 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:13:03.376263466 +0000 UTC m=+149.308600702 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.877990 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-trfnh"] Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.908698 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.914414 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.927896 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.977240 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.977455 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2788ab92-5c72-4f27-888e-ef695e4cf6cb-catalog-content\") pod \"community-operators-trfnh\" (UID: \"2788ab92-5c72-4f27-888e-ef695e4cf6cb\") " pod="openshift-marketplace/community-operators-trfnh" Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.977489 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2788ab92-5c72-4f27-888e-ef695e4cf6cb-utilities\") pod \"community-operators-trfnh\" (UID: \"2788ab92-5c72-4f27-888e-ef695e4cf6cb\") " pod="openshift-marketplace/community-operators-trfnh" Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.977549 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zc4gb\" (UniqueName: \"kubernetes.io/projected/2788ab92-5c72-4f27-888e-ef695e4cf6cb-kube-api-access-zc4gb\") pod \"community-operators-trfnh\" (UID: \"2788ab92-5c72-4f27-888e-ef695e4cf6cb\") " pod="openshift-marketplace/community-operators-trfnh" Oct 04 03:13:02 crc kubenswrapper[4742]: E1004 03:13:02.977977 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:13:03.477960941 +0000 UTC m=+149.410298107 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.978434 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2788ab92-5c72-4f27-888e-ef695e4cf6cb-catalog-content\") pod \"community-operators-trfnh\" (UID: \"2788ab92-5c72-4f27-888e-ef695e4cf6cb\") " pod="openshift-marketplace/community-operators-trfnh" Oct 04 03:13:02 crc kubenswrapper[4742]: I1004 03:13:02.978649 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2788ab92-5c72-4f27-888e-ef695e4cf6cb-utilities\") pod \"community-operators-trfnh\" (UID: \"2788ab92-5c72-4f27-888e-ef695e4cf6cb\") " pod="openshift-marketplace/community-operators-trfnh" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.037207 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zc4gb\" (UniqueName: \"kubernetes.io/projected/2788ab92-5c72-4f27-888e-ef695e4cf6cb-kube-api-access-zc4gb\") pod \"community-operators-trfnh\" (UID: \"2788ab92-5c72-4f27-888e-ef695e4cf6cb\") " pod="openshift-marketplace/community-operators-trfnh" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.048338 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-b5xzp"] Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.049721 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b5xzp" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.061663 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.067468 4742 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.073921 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b5xzp"] Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.079383 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcmvr\" (UniqueName: \"kubernetes.io/projected/f3b8a6f1-47db-4541-af94-5519b15d1407-kube-api-access-dcmvr\") pod \"certified-operators-b5xzp\" (UID: \"f3b8a6f1-47db-4541-af94-5519b15d1407\") " pod="openshift-marketplace/certified-operators-b5xzp" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.079435 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3b8a6f1-47db-4541-af94-5519b15d1407-catalog-content\") pod \"certified-operators-b5xzp\" (UID: \"f3b8a6f1-47db-4541-af94-5519b15d1407\") " pod="openshift-marketplace/certified-operators-b5xzp" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.079472 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.079513 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3b8a6f1-47db-4541-af94-5519b15d1407-utilities\") pod \"certified-operators-b5xzp\" (UID: \"f3b8a6f1-47db-4541-af94-5519b15d1407\") " pod="openshift-marketplace/certified-operators-b5xzp" Oct 04 03:13:03 crc kubenswrapper[4742]: E1004 03:13:03.079979 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:13:03.579957884 +0000 UTC m=+149.512295060 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.155964 4742 patch_prober.go:28] interesting pod/router-default-5444994796-85fhx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:13:03 crc kubenswrapper[4742]: [-]has-synced failed: reason withheld Oct 04 03:13:03 crc kubenswrapper[4742]: [+]process-running ok Oct 04 03:13:03 crc kubenswrapper[4742]: healthz check failed Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.156516 4742 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85fhx" podUID="093665e3-6dc0-4c51-a0f3-71363eef8367" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.167658 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-trfnh" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.184244 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:13:03 crc kubenswrapper[4742]: E1004 03:13:03.184454 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:13:03.684415323 +0000 UTC m=+149.616752499 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.184683 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcmvr\" (UniqueName: \"kubernetes.io/projected/f3b8a6f1-47db-4541-af94-5519b15d1407-kube-api-access-dcmvr\") pod \"certified-operators-b5xzp\" (UID: \"f3b8a6f1-47db-4541-af94-5519b15d1407\") " pod="openshift-marketplace/certified-operators-b5xzp" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.184728 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3b8a6f1-47db-4541-af94-5519b15d1407-catalog-content\") pod \"certified-operators-b5xzp\" (UID: \"f3b8a6f1-47db-4541-af94-5519b15d1407\") " pod="openshift-marketplace/certified-operators-b5xzp" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.184756 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.184805 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3b8a6f1-47db-4541-af94-5519b15d1407-utilities\") pod \"certified-operators-b5xzp\" (UID: \"f3b8a6f1-47db-4541-af94-5519b15d1407\") " pod="openshift-marketplace/certified-operators-b5xzp" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.185462 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3b8a6f1-47db-4541-af94-5519b15d1407-utilities\") pod \"certified-operators-b5xzp\" (UID: \"f3b8a6f1-47db-4541-af94-5519b15d1407\") " pod="openshift-marketplace/certified-operators-b5xzp" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.185704 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3b8a6f1-47db-4541-af94-5519b15d1407-catalog-content\") pod \"certified-operators-b5xzp\" (UID: \"f3b8a6f1-47db-4541-af94-5519b15d1407\") " pod="openshift-marketplace/certified-operators-b5xzp" Oct 04 03:13:03 crc kubenswrapper[4742]: E1004 03:13:03.186334 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:13:03.686313324 +0000 UTC m=+149.618650510 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.243548 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcmvr\" (UniqueName: \"kubernetes.io/projected/f3b8a6f1-47db-4541-af94-5519b15d1407-kube-api-access-dcmvr\") pod \"certified-operators-b5xzp\" (UID: \"f3b8a6f1-47db-4541-af94-5519b15d1407\") " pod="openshift-marketplace/certified-operators-b5xzp" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.248892 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-72m4v"] Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.249884 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-72m4v" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.270867 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-72m4v"] Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.299405 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.299732 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5f6j\" (UniqueName: \"kubernetes.io/projected/0fbf09fa-a448-4ac9-8550-e9c719ec22d2-kube-api-access-x5f6j\") pod \"community-operators-72m4v\" (UID: \"0fbf09fa-a448-4ac9-8550-e9c719ec22d2\") " pod="openshift-marketplace/community-operators-72m4v" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.299766 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fbf09fa-a448-4ac9-8550-e9c719ec22d2-catalog-content\") pod \"community-operators-72m4v\" (UID: \"0fbf09fa-a448-4ac9-8550-e9c719ec22d2\") " pod="openshift-marketplace/community-operators-72m4v" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.299788 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fbf09fa-a448-4ac9-8550-e9c719ec22d2-utilities\") pod \"community-operators-72m4v\" (UID: \"0fbf09fa-a448-4ac9-8550-e9c719ec22d2\") " pod="openshift-marketplace/community-operators-72m4v" Oct 04 03:13:03 crc kubenswrapper[4742]: E1004 03:13:03.299945 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:13:03.799922448 +0000 UTC m=+149.732259624 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.368823 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b5xzp" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.384294 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wn8jl" event={"ID":"35542dc5-96db-4eca-a2c2-85ebf65ea31f","Type":"ContainerStarted","Data":"98e096beba9a37c166077f811c42fb0a413fb296088846aaf444a4ac21c2dc62"} Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.415609 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5f6j\" (UniqueName: \"kubernetes.io/projected/0fbf09fa-a448-4ac9-8550-e9c719ec22d2-kube-api-access-x5f6j\") pod \"community-operators-72m4v\" (UID: \"0fbf09fa-a448-4ac9-8550-e9c719ec22d2\") " pod="openshift-marketplace/community-operators-72m4v" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.419831 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fbf09fa-a448-4ac9-8550-e9c719ec22d2-catalog-content\") pod \"community-operators-72m4v\" (UID: \"0fbf09fa-a448-4ac9-8550-e9c719ec22d2\") " pod="openshift-marketplace/community-operators-72m4v" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.420372 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fbf09fa-a448-4ac9-8550-e9c719ec22d2-utilities\") pod \"community-operators-72m4v\" (UID: \"0fbf09fa-a448-4ac9-8550-e9c719ec22d2\") " pod="openshift-marketplace/community-operators-72m4v" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.421043 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:13:03 crc kubenswrapper[4742]: E1004 03:13:03.423172 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:13:03.92315348 +0000 UTC m=+149.855490656 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.424319 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fbf09fa-a448-4ac9-8550-e9c719ec22d2-catalog-content\") pod \"community-operators-72m4v\" (UID: \"0fbf09fa-a448-4ac9-8550-e9c719ec22d2\") " pod="openshift-marketplace/community-operators-72m4v" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.424458 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fbf09fa-a448-4ac9-8550-e9c719ec22d2-utilities\") pod \"community-operators-72m4v\" (UID: \"0fbf09fa-a448-4ac9-8550-e9c719ec22d2\") " pod="openshift-marketplace/community-operators-72m4v" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.453238 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5f6j\" (UniqueName: \"kubernetes.io/projected/0fbf09fa-a448-4ac9-8550-e9c719ec22d2-kube-api-access-x5f6j\") pod \"community-operators-72m4v\" (UID: \"0fbf09fa-a448-4ac9-8550-e9c719ec22d2\") " pod="openshift-marketplace/community-operators-72m4v" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.484016 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-j5nft"] Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.486837 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-wn8jl" podStartSLOduration=10.486804636 podStartE2EDuration="10.486804636s" podCreationTimestamp="2025-10-04 03:12:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:13:03.448716275 +0000 UTC m=+149.381053451" watchObservedRunningTime="2025-10-04 03:13:03.486804636 +0000 UTC m=+149.419141822" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.494139 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j5nft" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.506327 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j5nft"] Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.526061 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:13:03 crc kubenswrapper[4742]: E1004 03:13:03.527641 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:13:04.027615089 +0000 UTC m=+149.959952265 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.584216 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-72m4v" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.627696 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9afa0f84-44ab-47b1-986f-b545fddb3824-catalog-content\") pod \"certified-operators-j5nft\" (UID: \"9afa0f84-44ab-47b1-986f-b545fddb3824\") " pod="openshift-marketplace/certified-operators-j5nft" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.627752 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9afa0f84-44ab-47b1-986f-b545fddb3824-utilities\") pod \"certified-operators-j5nft\" (UID: \"9afa0f84-44ab-47b1-986f-b545fddb3824\") " pod="openshift-marketplace/certified-operators-j5nft" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.627787 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.628047 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7l4mf\" (UniqueName: \"kubernetes.io/projected/9afa0f84-44ab-47b1-986f-b545fddb3824-kube-api-access-7l4mf\") pod \"certified-operators-j5nft\" (UID: \"9afa0f84-44ab-47b1-986f-b545fddb3824\") " pod="openshift-marketplace/certified-operators-j5nft" Oct 04 03:13:03 crc kubenswrapper[4742]: E1004 03:13:03.628512 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:13:04.128496563 +0000 UTC m=+150.060833739 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.730630 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:13:03 crc kubenswrapper[4742]: E1004 03:13:03.730723 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:13:04.2306498 +0000 UTC m=+150.162986976 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.731557 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9afa0f84-44ab-47b1-986f-b545fddb3824-catalog-content\") pod \"certified-operators-j5nft\" (UID: \"9afa0f84-44ab-47b1-986f-b545fddb3824\") " pod="openshift-marketplace/certified-operators-j5nft" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.731603 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9afa0f84-44ab-47b1-986f-b545fddb3824-utilities\") pod \"certified-operators-j5nft\" (UID: \"9afa0f84-44ab-47b1-986f-b545fddb3824\") " pod="openshift-marketplace/certified-operators-j5nft" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.731638 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.731697 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7l4mf\" (UniqueName: \"kubernetes.io/projected/9afa0f84-44ab-47b1-986f-b545fddb3824-kube-api-access-7l4mf\") pod \"certified-operators-j5nft\" (UID: \"9afa0f84-44ab-47b1-986f-b545fddb3824\") " pod="openshift-marketplace/certified-operators-j5nft" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.732752 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9afa0f84-44ab-47b1-986f-b545fddb3824-catalog-content\") pod \"certified-operators-j5nft\" (UID: \"9afa0f84-44ab-47b1-986f-b545fddb3824\") " pod="openshift-marketplace/certified-operators-j5nft" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.733052 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9afa0f84-44ab-47b1-986f-b545fddb3824-utilities\") pod \"certified-operators-j5nft\" (UID: \"9afa0f84-44ab-47b1-986f-b545fddb3824\") " pod="openshift-marketplace/certified-operators-j5nft" Oct 04 03:13:03 crc kubenswrapper[4742]: E1004 03:13:03.733413 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:13:04.233398334 +0000 UTC m=+150.165735510 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.766401 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7l4mf\" (UniqueName: \"kubernetes.io/projected/9afa0f84-44ab-47b1-986f-b545fddb3824-kube-api-access-7l4mf\") pod \"certified-operators-j5nft\" (UID: \"9afa0f84-44ab-47b1-986f-b545fddb3824\") " pod="openshift-marketplace/certified-operators-j5nft" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.833740 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:13:03 crc kubenswrapper[4742]: E1004 03:13:03.834248 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:13:04.334224055 +0000 UTC m=+150.266561231 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.837112 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-trfnh"] Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.872101 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b5xzp"] Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.910685 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j5nft" Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.936916 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:13:03 crc kubenswrapper[4742]: E1004 03:13:03.937295 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:13:04.437254746 +0000 UTC m=+150.369592102 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f6gcs" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:03 crc kubenswrapper[4742]: I1004 03:13:03.962830 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-72m4v"] Oct 04 03:13:03 crc kubenswrapper[4742]: W1004 03:13:03.964832 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-f8df94d494214c5f40a8d3294a0dc7c732fb6c034555e9447f32a64550ba20b5 WatchSource:0}: Error finding container f8df94d494214c5f40a8d3294a0dc7c732fb6c034555e9447f32a64550ba20b5: Status 404 returned error can't find the container with id f8df94d494214c5f40a8d3294a0dc7c732fb6c034555e9447f32a64550ba20b5 Oct 04 03:13:03 crc kubenswrapper[4742]: W1004 03:13:03.965501 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-21ac8897797f84fb2726ddf61dc20bae49778cd91ba9993ea0bdc2d249ac6c23 WatchSource:0}: Error finding container 21ac8897797f84fb2726ddf61dc20bae49778cd91ba9993ea0bdc2d249ac6c23: Status 404 returned error can't find the container with id 21ac8897797f84fb2726ddf61dc20bae49778cd91ba9993ea0bdc2d249ac6c23 Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.039786 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:13:04 crc kubenswrapper[4742]: E1004 03:13:04.040226 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:13:04.540207015 +0000 UTC m=+150.472544191 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.040261 4742 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-04T03:13:03.067506141Z","Handler":null,"Name":""} Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.045731 4742 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.045751 4742 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.131774 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.133668 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.137917 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.138224 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.140392 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.140973 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.144863 4742 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.144922 4742 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.152475 4742 patch_prober.go:28] interesting pod/router-default-5444994796-85fhx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:13:04 crc kubenswrapper[4742]: [-]has-synced failed: reason withheld Oct 04 03:13:04 crc kubenswrapper[4742]: [+]process-running ok Oct 04 03:13:04 crc kubenswrapper[4742]: healthz check failed Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.152532 4742 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85fhx" podUID="093665e3-6dc0-4c51-a0f3-71363eef8367" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.183790 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j5nft"] Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.190307 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f6gcs\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:13:04 crc kubenswrapper[4742]: W1004 03:13:04.190528 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9afa0f84_44ab_47b1_986f_b545fddb3824.slice/crio-eaee1d4a0823e5f19f946a1fab6da4d3821df5e36fd1854e8011973683532e18 WatchSource:0}: Error finding container eaee1d4a0823e5f19f946a1fab6da4d3821df5e36fd1854e8011973683532e18: Status 404 returned error can't find the container with id eaee1d4a0823e5f19f946a1fab6da4d3821df5e36fd1854e8011973683532e18 Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.220962 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.242198 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.243027 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/76ed571a-f997-4d52-8b5c-108211fe4da6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"76ed571a-f997-4d52-8b5c-108211fe4da6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.243100 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/76ed571a-f997-4d52-8b5c-108211fe4da6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"76ed571a-f997-4d52-8b5c-108211fe4da6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.248398 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.304051 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.304123 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.318259 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.350303 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/76ed571a-f997-4d52-8b5c-108211fe4da6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"76ed571a-f997-4d52-8b5c-108211fe4da6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.350389 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/76ed571a-f997-4d52-8b5c-108211fe4da6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"76ed571a-f997-4d52-8b5c-108211fe4da6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.350544 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/76ed571a-f997-4d52-8b5c-108211fe4da6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"76ed571a-f997-4d52-8b5c-108211fe4da6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.382007 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/76ed571a-f997-4d52-8b5c-108211fe4da6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"76ed571a-f997-4d52-8b5c-108211fe4da6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.401074 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"7c562d1f76a62b708d32dbcdca1fedb5170bd8a4164af2a02a8f092521e3e172"} Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.401158 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"835d471ad317002d5496678d3916c8446696105823a51c92f27b2988a5633cac"} Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.402112 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.414790 4742 generic.go:334] "Generic (PLEG): container finished" podID="2788ab92-5c72-4f27-888e-ef695e4cf6cb" containerID="1d902d8de38573af3480eb1b68d8d6853cb39f8197645dba851d1fecada4d514" exitCode=0 Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.415141 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-trfnh" event={"ID":"2788ab92-5c72-4f27-888e-ef695e4cf6cb","Type":"ContainerDied","Data":"1d902d8de38573af3480eb1b68d8d6853cb39f8197645dba851d1fecada4d514"} Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.415205 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-trfnh" event={"ID":"2788ab92-5c72-4f27-888e-ef695e4cf6cb","Type":"ContainerStarted","Data":"33984713aca33cb6b6070d96e5726eebb5378c76d2d78f8d66f09542b6ea47c0"} Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.421187 4742 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.440062 4742 generic.go:334] "Generic (PLEG): container finished" podID="7f11b3c3-6380-4f4d-9d03-622e66e33bc8" containerID="5da555775d193aac2a2c45269fd27c648394620f29687dc3df2a5b6c00baa45e" exitCode=0 Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.440179 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325780-9b22t" event={"ID":"7f11b3c3-6380-4f4d-9d03-622e66e33bc8","Type":"ContainerDied","Data":"5da555775d193aac2a2c45269fd27c648394620f29687dc3df2a5b6c00baa45e"} Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.450402 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j5nft" event={"ID":"9afa0f84-44ab-47b1-986f-b545fddb3824","Type":"ContainerStarted","Data":"eaee1d4a0823e5f19f946a1fab6da4d3821df5e36fd1854e8011973683532e18"} Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.455636 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.462656 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"7c788195234502e09b974d789bf9f874df2488ed8a60adddb4962f8433dd12e6"} Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.462719 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"21ac8897797f84fb2726ddf61dc20bae49778cd91ba9993ea0bdc2d249ac6c23"} Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.468409 4742 generic.go:334] "Generic (PLEG): container finished" podID="0fbf09fa-a448-4ac9-8550-e9c719ec22d2" containerID="26381f12e5108833d45d4c0aa97495c62d1f48d8f66c758baea0bf01127b3c03" exitCode=0 Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.468497 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72m4v" event={"ID":"0fbf09fa-a448-4ac9-8550-e9c719ec22d2","Type":"ContainerDied","Data":"26381f12e5108833d45d4c0aa97495c62d1f48d8f66c758baea0bf01127b3c03"} Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.468535 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72m4v" event={"ID":"0fbf09fa-a448-4ac9-8550-e9c719ec22d2","Type":"ContainerStarted","Data":"d72d75824473a8968c7b350ddf34e4f23193dad3c6bbd939045abbb9ad923fdd"} Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.471470 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"c0474286f6d59afff9cc91ea0976b6c02e3411636583e45cf66b2154b2a44594"} Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.471516 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"f8df94d494214c5f40a8d3294a0dc7c732fb6c034555e9447f32a64550ba20b5"} Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.482760 4742 generic.go:334] "Generic (PLEG): container finished" podID="f3b8a6f1-47db-4541-af94-5519b15d1407" containerID="452d6d3a8866d451b5f40bd252e13295fd55e9a672b02d533f288bd323d389eb" exitCode=0 Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.484228 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b5xzp" event={"ID":"f3b8a6f1-47db-4541-af94-5519b15d1407","Type":"ContainerDied","Data":"452d6d3a8866d451b5f40bd252e13295fd55e9a672b02d533f288bd323d389eb"} Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.485060 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b5xzp" event={"ID":"f3b8a6f1-47db-4541-af94-5519b15d1407","Type":"ContainerStarted","Data":"676af860861de4fc08f05d940e72e080eae8cb9624ba2cc6d276df0354743825"} Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.492934 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-f5978" Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.506440 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-f6gcs"] Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.834248 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lb54b"] Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.835947 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lb54b" Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.839846 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.856282 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lb54b"] Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.886262 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.965291 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3fa495f-e19d-4658-a335-1e49cb14a721-catalog-content\") pod \"redhat-marketplace-lb54b\" (UID: \"c3fa495f-e19d-4658-a335-1e49cb14a721\") " pod="openshift-marketplace/redhat-marketplace-lb54b" Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.965383 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7jjg\" (UniqueName: \"kubernetes.io/projected/c3fa495f-e19d-4658-a335-1e49cb14a721-kube-api-access-k7jjg\") pod \"redhat-marketplace-lb54b\" (UID: \"c3fa495f-e19d-4658-a335-1e49cb14a721\") " pod="openshift-marketplace/redhat-marketplace-lb54b" Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.965404 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3fa495f-e19d-4658-a335-1e49cb14a721-utilities\") pod \"redhat-marketplace-lb54b\" (UID: \"c3fa495f-e19d-4658-a335-1e49cb14a721\") " pod="openshift-marketplace/redhat-marketplace-lb54b" Oct 04 03:13:04 crc kubenswrapper[4742]: I1004 03:13:04.975486 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 04 03:13:04 crc kubenswrapper[4742]: W1004 03:13:04.989202 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod76ed571a_f997_4d52_8b5c_108211fe4da6.slice/crio-813a69551043a729ec8b581c9ae6821c89cce5d292e75147233b6ccf8ccd4ed0 WatchSource:0}: Error finding container 813a69551043a729ec8b581c9ae6821c89cce5d292e75147233b6ccf8ccd4ed0: Status 404 returned error can't find the container with id 813a69551043a729ec8b581c9ae6821c89cce5d292e75147233b6ccf8ccd4ed0 Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.066120 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8w75v" Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.067422 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3fa495f-e19d-4658-a335-1e49cb14a721-catalog-content\") pod \"redhat-marketplace-lb54b\" (UID: \"c3fa495f-e19d-4658-a335-1e49cb14a721\") " pod="openshift-marketplace/redhat-marketplace-lb54b" Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.067546 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7jjg\" (UniqueName: \"kubernetes.io/projected/c3fa495f-e19d-4658-a335-1e49cb14a721-kube-api-access-k7jjg\") pod \"redhat-marketplace-lb54b\" (UID: \"c3fa495f-e19d-4658-a335-1e49cb14a721\") " pod="openshift-marketplace/redhat-marketplace-lb54b" Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.067579 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3fa495f-e19d-4658-a335-1e49cb14a721-utilities\") pod \"redhat-marketplace-lb54b\" (UID: \"c3fa495f-e19d-4658-a335-1e49cb14a721\") " pod="openshift-marketplace/redhat-marketplace-lb54b" Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.068611 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3fa495f-e19d-4658-a335-1e49cb14a721-utilities\") pod \"redhat-marketplace-lb54b\" (UID: \"c3fa495f-e19d-4658-a335-1e49cb14a721\") " pod="openshift-marketplace/redhat-marketplace-lb54b" Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.068685 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3fa495f-e19d-4658-a335-1e49cb14a721-catalog-content\") pod \"redhat-marketplace-lb54b\" (UID: \"c3fa495f-e19d-4658-a335-1e49cb14a721\") " pod="openshift-marketplace/redhat-marketplace-lb54b" Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.124378 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7jjg\" (UniqueName: \"kubernetes.io/projected/c3fa495f-e19d-4658-a335-1e49cb14a721-kube-api-access-k7jjg\") pod \"redhat-marketplace-lb54b\" (UID: \"c3fa495f-e19d-4658-a335-1e49cb14a721\") " pod="openshift-marketplace/redhat-marketplace-lb54b" Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.153392 4742 patch_prober.go:28] interesting pod/router-default-5444994796-85fhx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:13:05 crc kubenswrapper[4742]: [-]has-synced failed: reason withheld Oct 04 03:13:05 crc kubenswrapper[4742]: [+]process-running ok Oct 04 03:13:05 crc kubenswrapper[4742]: healthz check failed Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.153508 4742 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85fhx" podUID="093665e3-6dc0-4c51-a0f3-71363eef8367" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.181910 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lb54b" Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.232002 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-46h54"] Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.233524 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-46h54" Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.247054 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-46h54"] Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.385815 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-ctkcp" Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.386611 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-ctkcp" Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.389205 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a-utilities\") pod \"redhat-marketplace-46h54\" (UID: \"7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a\") " pod="openshift-marketplace/redhat-marketplace-46h54" Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.389286 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a-catalog-content\") pod \"redhat-marketplace-46h54\" (UID: \"7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a\") " pod="openshift-marketplace/redhat-marketplace-46h54" Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.389361 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ssgl\" (UniqueName: \"kubernetes.io/projected/7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a-kube-api-access-4ssgl\") pod \"redhat-marketplace-46h54\" (UID: \"7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a\") " pod="openshift-marketplace/redhat-marketplace-46h54" Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.404482 4742 patch_prober.go:28] interesting pod/console-f9d7485db-ctkcp container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.404568 4742 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-ctkcp" podUID="5ba77307-01a4-43a3-8888-3da1c700ea2e" containerName="console" probeResult="failure" output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.434199 4742 patch_prober.go:28] interesting pod/downloads-7954f5f757-h22jn container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.434246 4742 patch_prober.go:28] interesting pod/downloads-7954f5f757-h22jn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.434295 4742 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-h22jn" podUID="0e5511f2-7587-4638-bd0c-af91c57a0163" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.434296 4742 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-h22jn" podUID="0e5511f2-7587-4638-bd0c-af91c57a0163" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.490679 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a-utilities\") pod \"redhat-marketplace-46h54\" (UID: \"7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a\") " pod="openshift-marketplace/redhat-marketplace-46h54" Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.490762 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a-catalog-content\") pod \"redhat-marketplace-46h54\" (UID: \"7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a\") " pod="openshift-marketplace/redhat-marketplace-46h54" Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.490810 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ssgl\" (UniqueName: \"kubernetes.io/projected/7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a-kube-api-access-4ssgl\") pod \"redhat-marketplace-46h54\" (UID: \"7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a\") " pod="openshift-marketplace/redhat-marketplace-46h54" Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.494080 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a-utilities\") pod \"redhat-marketplace-46h54\" (UID: \"7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a\") " pod="openshift-marketplace/redhat-marketplace-46h54" Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.495191 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a-catalog-content\") pod \"redhat-marketplace-46h54\" (UID: \"7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a\") " pod="openshift-marketplace/redhat-marketplace-46h54" Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.522616 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" event={"ID":"4892c4de-41a2-4c58-bb59-36a6c584da2f","Type":"ContainerStarted","Data":"8da953400f2531c9e7ce60b4d31e657c743b0c1b885feb997298d47edaf94623"} Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.522693 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" event={"ID":"4892c4de-41a2-4c58-bb59-36a6c584da2f","Type":"ContainerStarted","Data":"c410b0070a34c78d7e797e2b37e127fc41594f11f25c72577395a93be0a6cc03"} Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.522980 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.523350 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ssgl\" (UniqueName: \"kubernetes.io/projected/7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a-kube-api-access-4ssgl\") pod \"redhat-marketplace-46h54\" (UID: \"7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a\") " pod="openshift-marketplace/redhat-marketplace-46h54" Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.533620 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"76ed571a-f997-4d52-8b5c-108211fe4da6","Type":"ContainerStarted","Data":"813a69551043a729ec8b581c9ae6821c89cce5d292e75147233b6ccf8ccd4ed0"} Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.545243 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" podStartSLOduration=128.545223192 podStartE2EDuration="2m8.545223192s" podCreationTimestamp="2025-10-04 03:10:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:13:05.54291333 +0000 UTC m=+151.475250506" watchObservedRunningTime="2025-10-04 03:13:05.545223192 +0000 UTC m=+151.477560388" Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.573136 4742 generic.go:334] "Generic (PLEG): container finished" podID="9afa0f84-44ab-47b1-986f-b545fddb3824" containerID="4a2a982cee2ee0cf2fb0c8e8613b45a4b460695fc00f91d5cd3ae5a35f47305c" exitCode=0 Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.575102 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j5nft" event={"ID":"9afa0f84-44ab-47b1-986f-b545fddb3824","Type":"ContainerDied","Data":"4a2a982cee2ee0cf2fb0c8e8613b45a4b460695fc00f91d5cd3ae5a35f47305c"} Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.579371 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-46h54" Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.748344 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q" Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.748391 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q" Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.761469 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q" Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.818337 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lb54b"] Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.924803 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325780-9b22t" Oct 04 03:13:05 crc kubenswrapper[4742]: I1004 03:13:05.987710 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-46h54"] Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.014802 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7f11b3c3-6380-4f4d-9d03-622e66e33bc8-config-volume\") pod \"7f11b3c3-6380-4f4d-9d03-622e66e33bc8\" (UID: \"7f11b3c3-6380-4f4d-9d03-622e66e33bc8\") " Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.014900 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66ddt\" (UniqueName: \"kubernetes.io/projected/7f11b3c3-6380-4f4d-9d03-622e66e33bc8-kube-api-access-66ddt\") pod \"7f11b3c3-6380-4f4d-9d03-622e66e33bc8\" (UID: \"7f11b3c3-6380-4f4d-9d03-622e66e33bc8\") " Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.015009 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7f11b3c3-6380-4f4d-9d03-622e66e33bc8-secret-volume\") pod \"7f11b3c3-6380-4f4d-9d03-622e66e33bc8\" (UID: \"7f11b3c3-6380-4f4d-9d03-622e66e33bc8\") " Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.016774 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f11b3c3-6380-4f4d-9d03-622e66e33bc8-config-volume" (OuterVolumeSpecName: "config-volume") pod "7f11b3c3-6380-4f4d-9d03-622e66e33bc8" (UID: "7f11b3c3-6380-4f4d-9d03-622e66e33bc8"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.027469 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f11b3c3-6380-4f4d-9d03-622e66e33bc8-kube-api-access-66ddt" (OuterVolumeSpecName: "kube-api-access-66ddt") pod "7f11b3c3-6380-4f4d-9d03-622e66e33bc8" (UID: "7f11b3c3-6380-4f4d-9d03-622e66e33bc8"). InnerVolumeSpecName "kube-api-access-66ddt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.039699 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f11b3c3-6380-4f4d-9d03-622e66e33bc8-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7f11b3c3-6380-4f4d-9d03-622e66e33bc8" (UID: "7f11b3c3-6380-4f4d-9d03-622e66e33bc8"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.117496 4742 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7f11b3c3-6380-4f4d-9d03-622e66e33bc8-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.117671 4742 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7f11b3c3-6380-4f4d-9d03-622e66e33bc8-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.117689 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66ddt\" (UniqueName: \"kubernetes.io/projected/7f11b3c3-6380-4f4d-9d03-622e66e33bc8-kube-api-access-66ddt\") on node \"crc\" DevicePath \"\"" Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.144864 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-85fhx" Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.150357 4742 patch_prober.go:28] interesting pod/router-default-5444994796-85fhx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:13:06 crc kubenswrapper[4742]: [-]has-synced failed: reason withheld Oct 04 03:13:06 crc kubenswrapper[4742]: [+]process-running ok Oct 04 03:13:06 crc kubenswrapper[4742]: healthz check failed Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.150401 4742 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85fhx" podUID="093665e3-6dc0-4c51-a0f3-71363eef8367" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.175207 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-tfc7g" Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.243883 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-67ztt"] Oct 04 03:13:06 crc kubenswrapper[4742]: E1004 03:13:06.244213 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f11b3c3-6380-4f4d-9d03-622e66e33bc8" containerName="collect-profiles" Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.244228 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f11b3c3-6380-4f4d-9d03-622e66e33bc8" containerName="collect-profiles" Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.244410 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f11b3c3-6380-4f4d-9d03-622e66e33bc8" containerName="collect-profiles" Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.245220 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-67ztt" Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.247214 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.260604 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-67ztt"] Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.432947 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c210acab-8945-4f9d-a8fe-47ae7d78a064-utilities\") pod \"redhat-operators-67ztt\" (UID: \"c210acab-8945-4f9d-a8fe-47ae7d78a064\") " pod="openshift-marketplace/redhat-operators-67ztt" Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.433053 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nj7w\" (UniqueName: \"kubernetes.io/projected/c210acab-8945-4f9d-a8fe-47ae7d78a064-kube-api-access-2nj7w\") pod \"redhat-operators-67ztt\" (UID: \"c210acab-8945-4f9d-a8fe-47ae7d78a064\") " pod="openshift-marketplace/redhat-operators-67ztt" Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.433087 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c210acab-8945-4f9d-a8fe-47ae7d78a064-catalog-content\") pod \"redhat-operators-67ztt\" (UID: \"c210acab-8945-4f9d-a8fe-47ae7d78a064\") " pod="openshift-marketplace/redhat-operators-67ztt" Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.536914 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c210acab-8945-4f9d-a8fe-47ae7d78a064-utilities\") pod \"redhat-operators-67ztt\" (UID: \"c210acab-8945-4f9d-a8fe-47ae7d78a064\") " pod="openshift-marketplace/redhat-operators-67ztt" Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.537004 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nj7w\" (UniqueName: \"kubernetes.io/projected/c210acab-8945-4f9d-a8fe-47ae7d78a064-kube-api-access-2nj7w\") pod \"redhat-operators-67ztt\" (UID: \"c210acab-8945-4f9d-a8fe-47ae7d78a064\") " pod="openshift-marketplace/redhat-operators-67ztt" Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.537023 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c210acab-8945-4f9d-a8fe-47ae7d78a064-catalog-content\") pod \"redhat-operators-67ztt\" (UID: \"c210acab-8945-4f9d-a8fe-47ae7d78a064\") " pod="openshift-marketplace/redhat-operators-67ztt" Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.538893 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c210acab-8945-4f9d-a8fe-47ae7d78a064-utilities\") pod \"redhat-operators-67ztt\" (UID: \"c210acab-8945-4f9d-a8fe-47ae7d78a064\") " pod="openshift-marketplace/redhat-operators-67ztt" Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.541746 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c210acab-8945-4f9d-a8fe-47ae7d78a064-catalog-content\") pod \"redhat-operators-67ztt\" (UID: \"c210acab-8945-4f9d-a8fe-47ae7d78a064\") " pod="openshift-marketplace/redhat-operators-67ztt" Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.571457 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nj7w\" (UniqueName: \"kubernetes.io/projected/c210acab-8945-4f9d-a8fe-47ae7d78a064-kube-api-access-2nj7w\") pod \"redhat-operators-67ztt\" (UID: \"c210acab-8945-4f9d-a8fe-47ae7d78a064\") " pod="openshift-marketplace/redhat-operators-67ztt" Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.585644 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-67ztt" Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.588007 4742 generic.go:334] "Generic (PLEG): container finished" podID="7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a" containerID="fc957861f4ebcf5836dbed163010972ff837169e0506a8b5877af8849336c973" exitCode=0 Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.588097 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-46h54" event={"ID":"7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a","Type":"ContainerDied","Data":"fc957861f4ebcf5836dbed163010972ff837169e0506a8b5877af8849336c973"} Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.588132 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-46h54" event={"ID":"7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a","Type":"ContainerStarted","Data":"835a760f9d53c3a1034dff098d37b9e473eaa73674cf462fdfa40d3db20d0559"} Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.605324 4742 generic.go:334] "Generic (PLEG): container finished" podID="76ed571a-f997-4d52-8b5c-108211fe4da6" containerID="e1898d0697111f8fcd605f3da1f89eafd26291aa5347ab82e39c6e94f42487cd" exitCode=0 Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.605408 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"76ed571a-f997-4d52-8b5c-108211fe4da6","Type":"ContainerDied","Data":"e1898d0697111f8fcd605f3da1f89eafd26291aa5347ab82e39c6e94f42487cd"} Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.611779 4742 generic.go:334] "Generic (PLEG): container finished" podID="c3fa495f-e19d-4658-a335-1e49cb14a721" containerID="c1283f46e50e62bf8c7be17cc716c1b39067f93311f464b7dcc25720c0655e97" exitCode=0 Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.611836 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lb54b" event={"ID":"c3fa495f-e19d-4658-a335-1e49cb14a721","Type":"ContainerDied","Data":"c1283f46e50e62bf8c7be17cc716c1b39067f93311f464b7dcc25720c0655e97"} Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.611861 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lb54b" event={"ID":"c3fa495f-e19d-4658-a335-1e49cb14a721","Type":"ContainerStarted","Data":"db7fa3e13b3a8fa3e4ca52d2b1ff0eaf0ac9d4bb82ed7cd2fa3f6ea7e46eb2b4"} Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.615639 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325780-9b22t" event={"ID":"7f11b3c3-6380-4f4d-9d03-622e66e33bc8","Type":"ContainerDied","Data":"f316ca86005256022008feff320c8d58a736775c9530c7176468953d9853697a"} Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.615667 4742 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f316ca86005256022008feff320c8d58a736775c9530c7176468953d9853697a" Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.616701 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325780-9b22t" Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.639298 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-mrq8q" Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.671404 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rd5tv"] Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.673155 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rd5tv" Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.680777 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rd5tv"] Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.849899 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3848a0f4-0c53-4a32-ad20-d4bce94d9a95-utilities\") pod \"redhat-operators-rd5tv\" (UID: \"3848a0f4-0c53-4a32-ad20-d4bce94d9a95\") " pod="openshift-marketplace/redhat-operators-rd5tv" Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.850395 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzc98\" (UniqueName: \"kubernetes.io/projected/3848a0f4-0c53-4a32-ad20-d4bce94d9a95-kube-api-access-gzc98\") pod \"redhat-operators-rd5tv\" (UID: \"3848a0f4-0c53-4a32-ad20-d4bce94d9a95\") " pod="openshift-marketplace/redhat-operators-rd5tv" Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.850429 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3848a0f4-0c53-4a32-ad20-d4bce94d9a95-catalog-content\") pod \"redhat-operators-rd5tv\" (UID: \"3848a0f4-0c53-4a32-ad20-d4bce94d9a95\") " pod="openshift-marketplace/redhat-operators-rd5tv" Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.952784 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3848a0f4-0c53-4a32-ad20-d4bce94d9a95-utilities\") pod \"redhat-operators-rd5tv\" (UID: \"3848a0f4-0c53-4a32-ad20-d4bce94d9a95\") " pod="openshift-marketplace/redhat-operators-rd5tv" Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.952846 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzc98\" (UniqueName: \"kubernetes.io/projected/3848a0f4-0c53-4a32-ad20-d4bce94d9a95-kube-api-access-gzc98\") pod \"redhat-operators-rd5tv\" (UID: \"3848a0f4-0c53-4a32-ad20-d4bce94d9a95\") " pod="openshift-marketplace/redhat-operators-rd5tv" Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.952894 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3848a0f4-0c53-4a32-ad20-d4bce94d9a95-catalog-content\") pod \"redhat-operators-rd5tv\" (UID: \"3848a0f4-0c53-4a32-ad20-d4bce94d9a95\") " pod="openshift-marketplace/redhat-operators-rd5tv" Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.953557 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3848a0f4-0c53-4a32-ad20-d4bce94d9a95-catalog-content\") pod \"redhat-operators-rd5tv\" (UID: \"3848a0f4-0c53-4a32-ad20-d4bce94d9a95\") " pod="openshift-marketplace/redhat-operators-rd5tv" Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.953686 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3848a0f4-0c53-4a32-ad20-d4bce94d9a95-utilities\") pod \"redhat-operators-rd5tv\" (UID: \"3848a0f4-0c53-4a32-ad20-d4bce94d9a95\") " pod="openshift-marketplace/redhat-operators-rd5tv" Oct 04 03:13:06 crc kubenswrapper[4742]: I1004 03:13:06.981093 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzc98\" (UniqueName: \"kubernetes.io/projected/3848a0f4-0c53-4a32-ad20-d4bce94d9a95-kube-api-access-gzc98\") pod \"redhat-operators-rd5tv\" (UID: \"3848a0f4-0c53-4a32-ad20-d4bce94d9a95\") " pod="openshift-marketplace/redhat-operators-rd5tv" Oct 04 03:13:07 crc kubenswrapper[4742]: I1004 03:13:07.021615 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rd5tv" Oct 04 03:13:07 crc kubenswrapper[4742]: I1004 03:13:07.105753 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-67ztt"] Oct 04 03:13:07 crc kubenswrapper[4742]: W1004 03:13:07.141537 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc210acab_8945_4f9d_a8fe_47ae7d78a064.slice/crio-eb25dc8b6ffd2d353caddfd5fe77515a24b7fe51a4ba4beab8250b748b804d9c WatchSource:0}: Error finding container eb25dc8b6ffd2d353caddfd5fe77515a24b7fe51a4ba4beab8250b748b804d9c: Status 404 returned error can't find the container with id eb25dc8b6ffd2d353caddfd5fe77515a24b7fe51a4ba4beab8250b748b804d9c Oct 04 03:13:07 crc kubenswrapper[4742]: I1004 03:13:07.148903 4742 patch_prober.go:28] interesting pod/router-default-5444994796-85fhx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:13:07 crc kubenswrapper[4742]: [-]has-synced failed: reason withheld Oct 04 03:13:07 crc kubenswrapper[4742]: [+]process-running ok Oct 04 03:13:07 crc kubenswrapper[4742]: healthz check failed Oct 04 03:13:07 crc kubenswrapper[4742]: I1004 03:13:07.149031 4742 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85fhx" podUID="093665e3-6dc0-4c51-a0f3-71363eef8367" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:13:07 crc kubenswrapper[4742]: I1004 03:13:07.557192 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rd5tv"] Oct 04 03:13:07 crc kubenswrapper[4742]: W1004 03:13:07.621786 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3848a0f4_0c53_4a32_ad20_d4bce94d9a95.slice/crio-f6535f2b3572d9958334a113f49ef0c9e2abb7419f6d5032e5041a4f5b7b9ebc WatchSource:0}: Error finding container f6535f2b3572d9958334a113f49ef0c9e2abb7419f6d5032e5041a4f5b7b9ebc: Status 404 returned error can't find the container with id f6535f2b3572d9958334a113f49ef0c9e2abb7419f6d5032e5041a4f5b7b9ebc Oct 04 03:13:07 crc kubenswrapper[4742]: I1004 03:13:07.636802 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-67ztt" event={"ID":"c210acab-8945-4f9d-a8fe-47ae7d78a064","Type":"ContainerDied","Data":"4a8c0ccaa12e14dfe6317a9f6c00ea1d59790e8a58ed8c1b29595dc47b51e8e3"} Oct 04 03:13:07 crc kubenswrapper[4742]: I1004 03:13:07.636759 4742 generic.go:334] "Generic (PLEG): container finished" podID="c210acab-8945-4f9d-a8fe-47ae7d78a064" containerID="4a8c0ccaa12e14dfe6317a9f6c00ea1d59790e8a58ed8c1b29595dc47b51e8e3" exitCode=0 Oct 04 03:13:07 crc kubenswrapper[4742]: I1004 03:13:07.636996 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-67ztt" event={"ID":"c210acab-8945-4f9d-a8fe-47ae7d78a064","Type":"ContainerStarted","Data":"eb25dc8b6ffd2d353caddfd5fe77515a24b7fe51a4ba4beab8250b748b804d9c"} Oct 04 03:13:08 crc kubenswrapper[4742]: I1004 03:13:08.036618 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 03:13:08 crc kubenswrapper[4742]: I1004 03:13:08.147992 4742 patch_prober.go:28] interesting pod/router-default-5444994796-85fhx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:13:08 crc kubenswrapper[4742]: [-]has-synced failed: reason withheld Oct 04 03:13:08 crc kubenswrapper[4742]: [+]process-running ok Oct 04 03:13:08 crc kubenswrapper[4742]: healthz check failed Oct 04 03:13:08 crc kubenswrapper[4742]: I1004 03:13:08.148083 4742 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85fhx" podUID="093665e3-6dc0-4c51-a0f3-71363eef8367" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:13:08 crc kubenswrapper[4742]: I1004 03:13:08.179780 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/76ed571a-f997-4d52-8b5c-108211fe4da6-kubelet-dir\") pod \"76ed571a-f997-4d52-8b5c-108211fe4da6\" (UID: \"76ed571a-f997-4d52-8b5c-108211fe4da6\") " Oct 04 03:13:08 crc kubenswrapper[4742]: I1004 03:13:08.180053 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/76ed571a-f997-4d52-8b5c-108211fe4da6-kube-api-access\") pod \"76ed571a-f997-4d52-8b5c-108211fe4da6\" (UID: \"76ed571a-f997-4d52-8b5c-108211fe4da6\") " Oct 04 03:13:08 crc kubenswrapper[4742]: I1004 03:13:08.180264 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/76ed571a-f997-4d52-8b5c-108211fe4da6-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "76ed571a-f997-4d52-8b5c-108211fe4da6" (UID: "76ed571a-f997-4d52-8b5c-108211fe4da6"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:13:08 crc kubenswrapper[4742]: I1004 03:13:08.180845 4742 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/76ed571a-f997-4d52-8b5c-108211fe4da6-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 04 03:13:08 crc kubenswrapper[4742]: I1004 03:13:08.240347 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76ed571a-f997-4d52-8b5c-108211fe4da6-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "76ed571a-f997-4d52-8b5c-108211fe4da6" (UID: "76ed571a-f997-4d52-8b5c-108211fe4da6"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:13:08 crc kubenswrapper[4742]: I1004 03:13:08.282030 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/76ed571a-f997-4d52-8b5c-108211fe4da6-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 03:13:08 crc kubenswrapper[4742]: I1004 03:13:08.662181 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rd5tv" event={"ID":"3848a0f4-0c53-4a32-ad20-d4bce94d9a95","Type":"ContainerStarted","Data":"f6535f2b3572d9958334a113f49ef0c9e2abb7419f6d5032e5041a4f5b7b9ebc"} Oct 04 03:13:08 crc kubenswrapper[4742]: I1004 03:13:08.666157 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"76ed571a-f997-4d52-8b5c-108211fe4da6","Type":"ContainerDied","Data":"813a69551043a729ec8b581c9ae6821c89cce5d292e75147233b6ccf8ccd4ed0"} Oct 04 03:13:08 crc kubenswrapper[4742]: I1004 03:13:08.666215 4742 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="813a69551043a729ec8b581c9ae6821c89cce5d292e75147233b6ccf8ccd4ed0" Oct 04 03:13:08 crc kubenswrapper[4742]: I1004 03:13:08.666318 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 03:13:09 crc kubenswrapper[4742]: I1004 03:13:09.156069 4742 patch_prober.go:28] interesting pod/router-default-5444994796-85fhx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:13:09 crc kubenswrapper[4742]: [-]has-synced failed: reason withheld Oct 04 03:13:09 crc kubenswrapper[4742]: [+]process-running ok Oct 04 03:13:09 crc kubenswrapper[4742]: healthz check failed Oct 04 03:13:09 crc kubenswrapper[4742]: I1004 03:13:09.157557 4742 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85fhx" podUID="093665e3-6dc0-4c51-a0f3-71363eef8367" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:13:09 crc kubenswrapper[4742]: I1004 03:13:09.184373 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 04 03:13:09 crc kubenswrapper[4742]: E1004 03:13:09.184639 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76ed571a-f997-4d52-8b5c-108211fe4da6" containerName="pruner" Oct 04 03:13:09 crc kubenswrapper[4742]: I1004 03:13:09.184653 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="76ed571a-f997-4d52-8b5c-108211fe4da6" containerName="pruner" Oct 04 03:13:09 crc kubenswrapper[4742]: I1004 03:13:09.184787 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="76ed571a-f997-4d52-8b5c-108211fe4da6" containerName="pruner" Oct 04 03:13:09 crc kubenswrapper[4742]: I1004 03:13:09.185231 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 03:13:09 crc kubenswrapper[4742]: I1004 03:13:09.189994 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 04 03:13:09 crc kubenswrapper[4742]: I1004 03:13:09.190836 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 04 03:13:09 crc kubenswrapper[4742]: I1004 03:13:09.193045 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 04 03:13:09 crc kubenswrapper[4742]: I1004 03:13:09.308871 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/afe664a2-9cfc-4a72-9903-08b2dbba5d3c-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"afe664a2-9cfc-4a72-9903-08b2dbba5d3c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 03:13:09 crc kubenswrapper[4742]: I1004 03:13:09.308950 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/afe664a2-9cfc-4a72-9903-08b2dbba5d3c-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"afe664a2-9cfc-4a72-9903-08b2dbba5d3c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 03:13:09 crc kubenswrapper[4742]: I1004 03:13:09.410148 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/afe664a2-9cfc-4a72-9903-08b2dbba5d3c-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"afe664a2-9cfc-4a72-9903-08b2dbba5d3c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 03:13:09 crc kubenswrapper[4742]: I1004 03:13:09.410218 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/afe664a2-9cfc-4a72-9903-08b2dbba5d3c-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"afe664a2-9cfc-4a72-9903-08b2dbba5d3c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 03:13:09 crc kubenswrapper[4742]: I1004 03:13:09.410810 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/afe664a2-9cfc-4a72-9903-08b2dbba5d3c-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"afe664a2-9cfc-4a72-9903-08b2dbba5d3c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 03:13:09 crc kubenswrapper[4742]: I1004 03:13:09.540509 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/afe664a2-9cfc-4a72-9903-08b2dbba5d3c-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"afe664a2-9cfc-4a72-9903-08b2dbba5d3c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 03:13:09 crc kubenswrapper[4742]: I1004 03:13:09.691469 4742 generic.go:334] "Generic (PLEG): container finished" podID="3848a0f4-0c53-4a32-ad20-d4bce94d9a95" containerID="1eab366f2d8201e823eab2977bab4efc5034f1efa0376b8a0a06e2c7ccbcb184" exitCode=0 Oct 04 03:13:09 crc kubenswrapper[4742]: I1004 03:13:09.691537 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rd5tv" event={"ID":"3848a0f4-0c53-4a32-ad20-d4bce94d9a95","Type":"ContainerDied","Data":"1eab366f2d8201e823eab2977bab4efc5034f1efa0376b8a0a06e2c7ccbcb184"} Oct 04 03:13:09 crc kubenswrapper[4742]: I1004 03:13:09.829472 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 03:13:10 crc kubenswrapper[4742]: I1004 03:13:10.152777 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-85fhx" Oct 04 03:13:10 crc kubenswrapper[4742]: I1004 03:13:10.155900 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-85fhx" Oct 04 03:13:10 crc kubenswrapper[4742]: I1004 03:13:10.219530 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 04 03:13:10 crc kubenswrapper[4742]: I1004 03:13:10.750853 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"afe664a2-9cfc-4a72-9903-08b2dbba5d3c","Type":"ContainerStarted","Data":"f5b132dd2284324c9ac601c474993210e310c8c3f7a86848f123c1e1b69f98c6"} Oct 04 03:13:10 crc kubenswrapper[4742]: I1004 03:13:10.999368 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-nvt5z" Oct 04 03:13:11 crc kubenswrapper[4742]: I1004 03:13:11.763874 4742 generic.go:334] "Generic (PLEG): container finished" podID="afe664a2-9cfc-4a72-9903-08b2dbba5d3c" containerID="f359111cad5ca5550e5ab7691c986182afbedc73f5e801315eb59ac46efe7557" exitCode=0 Oct 04 03:13:11 crc kubenswrapper[4742]: I1004 03:13:11.764303 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"afe664a2-9cfc-4a72-9903-08b2dbba5d3c","Type":"ContainerDied","Data":"f359111cad5ca5550e5ab7691c986182afbedc73f5e801315eb59ac46efe7557"} Oct 04 03:13:14 crc kubenswrapper[4742]: I1004 03:13:14.871895 4742 patch_prober.go:28] interesting pod/machine-config-daemon-qdt7v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:13:14 crc kubenswrapper[4742]: I1004 03:13:14.872531 4742 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:13:15 crc kubenswrapper[4742]: I1004 03:13:15.431402 4742 patch_prober.go:28] interesting pod/downloads-7954f5f757-h22jn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Oct 04 03:13:15 crc kubenswrapper[4742]: I1004 03:13:15.431462 4742 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-h22jn" podUID="0e5511f2-7587-4638-bd0c-af91c57a0163" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Oct 04 03:13:15 crc kubenswrapper[4742]: I1004 03:13:15.431476 4742 patch_prober.go:28] interesting pod/downloads-7954f5f757-h22jn container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Oct 04 03:13:15 crc kubenswrapper[4742]: I1004 03:13:15.431529 4742 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-h22jn" podUID="0e5511f2-7587-4638-bd0c-af91c57a0163" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Oct 04 03:13:15 crc kubenswrapper[4742]: I1004 03:13:15.451192 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-ctkcp" Oct 04 03:13:15 crc kubenswrapper[4742]: I1004 03:13:15.460759 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-ctkcp" Oct 04 03:13:18 crc kubenswrapper[4742]: I1004 03:13:18.421132 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 03:13:18 crc kubenswrapper[4742]: I1004 03:13:18.583847 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/afe664a2-9cfc-4a72-9903-08b2dbba5d3c-kube-api-access\") pod \"afe664a2-9cfc-4a72-9903-08b2dbba5d3c\" (UID: \"afe664a2-9cfc-4a72-9903-08b2dbba5d3c\") " Oct 04 03:13:18 crc kubenswrapper[4742]: I1004 03:13:18.584039 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/afe664a2-9cfc-4a72-9903-08b2dbba5d3c-kubelet-dir\") pod \"afe664a2-9cfc-4a72-9903-08b2dbba5d3c\" (UID: \"afe664a2-9cfc-4a72-9903-08b2dbba5d3c\") " Oct 04 03:13:18 crc kubenswrapper[4742]: I1004 03:13:18.584181 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/afe664a2-9cfc-4a72-9903-08b2dbba5d3c-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "afe664a2-9cfc-4a72-9903-08b2dbba5d3c" (UID: "afe664a2-9cfc-4a72-9903-08b2dbba5d3c"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:13:18 crc kubenswrapper[4742]: I1004 03:13:18.594769 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afe664a2-9cfc-4a72-9903-08b2dbba5d3c-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "afe664a2-9cfc-4a72-9903-08b2dbba5d3c" (UID: "afe664a2-9cfc-4a72-9903-08b2dbba5d3c"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:13:18 crc kubenswrapper[4742]: I1004 03:13:18.685871 4742 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/afe664a2-9cfc-4a72-9903-08b2dbba5d3c-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 04 03:13:18 crc kubenswrapper[4742]: I1004 03:13:18.685909 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/afe664a2-9cfc-4a72-9903-08b2dbba5d3c-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 03:13:18 crc kubenswrapper[4742]: I1004 03:13:18.828185 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"afe664a2-9cfc-4a72-9903-08b2dbba5d3c","Type":"ContainerDied","Data":"f5b132dd2284324c9ac601c474993210e310c8c3f7a86848f123c1e1b69f98c6"} Oct 04 03:13:18 crc kubenswrapper[4742]: I1004 03:13:18.828221 4742 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f5b132dd2284324c9ac601c474993210e310c8c3f7a86848f123c1e1b69f98c6" Oct 04 03:13:18 crc kubenswrapper[4742]: I1004 03:13:18.828309 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 03:13:20 crc kubenswrapper[4742]: I1004 03:13:20.714319 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0302950c-deb3-420a-b798-c99459e62b3f-metrics-certs\") pod \"network-metrics-daemon-cd46x\" (UID: \"0302950c-deb3-420a-b798-c99459e62b3f\") " pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:13:20 crc kubenswrapper[4742]: I1004 03:13:20.904307 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0302950c-deb3-420a-b798-c99459e62b3f-metrics-certs\") pod \"network-metrics-daemon-cd46x\" (UID: \"0302950c-deb3-420a-b798-c99459e62b3f\") " pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:13:21 crc kubenswrapper[4742]: I1004 03:13:21.186576 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cd46x" Oct 04 03:13:24 crc kubenswrapper[4742]: I1004 03:13:24.226940 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:13:25 crc kubenswrapper[4742]: I1004 03:13:25.438462 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-h22jn" Oct 04 03:13:35 crc kubenswrapper[4742]: I1004 03:13:35.923790 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xg5l9" Oct 04 03:13:39 crc kubenswrapper[4742]: E1004 03:13:39.670245 4742 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 04 03:13:39 crc kubenswrapper[4742]: E1004 03:13:39.670702 4742 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zc4gb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-trfnh_openshift-marketplace(2788ab92-5c72-4f27-888e-ef695e4cf6cb): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 03:13:39 crc kubenswrapper[4742]: E1004 03:13:39.671867 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-trfnh" podUID="2788ab92-5c72-4f27-888e-ef695e4cf6cb" Oct 04 03:13:41 crc kubenswrapper[4742]: E1004 03:13:41.214757 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-trfnh" podUID="2788ab92-5c72-4f27-888e-ef695e4cf6cb" Oct 04 03:13:41 crc kubenswrapper[4742]: E1004 03:13:41.286291 4742 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 04 03:13:41 crc kubenswrapper[4742]: E1004 03:13:41.286446 4742 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7l4mf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-j5nft_openshift-marketplace(9afa0f84-44ab-47b1-986f-b545fddb3824): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 03:13:41 crc kubenswrapper[4742]: E1004 03:13:41.287545 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-j5nft" podUID="9afa0f84-44ab-47b1-986f-b545fddb3824" Oct 04 03:13:41 crc kubenswrapper[4742]: E1004 03:13:41.309727 4742 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 04 03:13:41 crc kubenswrapper[4742]: E1004 03:13:41.309900 4742 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x5f6j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-72m4v_openshift-marketplace(0fbf09fa-a448-4ac9-8550-e9c719ec22d2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 03:13:41 crc kubenswrapper[4742]: E1004 03:13:41.311099 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-72m4v" podUID="0fbf09fa-a448-4ac9-8550-e9c719ec22d2" Oct 04 03:13:42 crc kubenswrapper[4742]: I1004 03:13:42.915638 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:13:43 crc kubenswrapper[4742]: E1004 03:13:43.991672 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-72m4v" podUID="0fbf09fa-a448-4ac9-8550-e9c719ec22d2" Oct 04 03:13:43 crc kubenswrapper[4742]: E1004 03:13:43.991676 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-j5nft" podUID="9afa0f84-44ab-47b1-986f-b545fddb3824" Oct 04 03:13:44 crc kubenswrapper[4742]: E1004 03:13:44.059672 4742 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 04 03:13:44 crc kubenswrapper[4742]: E1004 03:13:44.059906 4742 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gzc98,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-rd5tv_openshift-marketplace(3848a0f4-0c53-4a32-ad20-d4bce94d9a95): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 03:13:44 crc kubenswrapper[4742]: E1004 03:13:44.061146 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-rd5tv" podUID="3848a0f4-0c53-4a32-ad20-d4bce94d9a95" Oct 04 03:13:44 crc kubenswrapper[4742]: I1004 03:13:44.871787 4742 patch_prober.go:28] interesting pod/machine-config-daemon-qdt7v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:13:44 crc kubenswrapper[4742]: I1004 03:13:44.872135 4742 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:13:46 crc kubenswrapper[4742]: E1004 03:13:46.428429 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-rd5tv" podUID="3848a0f4-0c53-4a32-ad20-d4bce94d9a95" Oct 04 03:13:46 crc kubenswrapper[4742]: E1004 03:13:46.507788 4742 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 04 03:13:46 crc kubenswrapper[4742]: E1004 03:13:46.507966 4742 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dcmvr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-b5xzp_openshift-marketplace(f3b8a6f1-47db-4541-af94-5519b15d1407): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 03:13:46 crc kubenswrapper[4742]: E1004 03:13:46.509162 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-b5xzp" podUID="f3b8a6f1-47db-4541-af94-5519b15d1407" Oct 04 03:13:46 crc kubenswrapper[4742]: E1004 03:13:46.553090 4742 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 04 03:13:46 crc kubenswrapper[4742]: E1004 03:13:46.553302 4742 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2nj7w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-67ztt_openshift-marketplace(c210acab-8945-4f9d-a8fe-47ae7d78a064): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 03:13:46 crc kubenswrapper[4742]: E1004 03:13:46.554466 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-67ztt" podUID="c210acab-8945-4f9d-a8fe-47ae7d78a064" Oct 04 03:13:46 crc kubenswrapper[4742]: E1004 03:13:46.971637 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-b5xzp" podUID="f3b8a6f1-47db-4541-af94-5519b15d1407" Oct 04 03:13:46 crc kubenswrapper[4742]: E1004 03:13:46.971644 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-67ztt" podUID="c210acab-8945-4f9d-a8fe-47ae7d78a064" Oct 04 03:13:47 crc kubenswrapper[4742]: E1004 03:13:47.047388 4742 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 04 03:13:47 crc kubenswrapper[4742]: E1004 03:13:47.047558 4742 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k7jjg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-lb54b_openshift-marketplace(c3fa495f-e19d-4658-a335-1e49cb14a721): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 03:13:47 crc kubenswrapper[4742]: E1004 03:13:47.048908 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-lb54b" podUID="c3fa495f-e19d-4658-a335-1e49cb14a721" Oct 04 03:13:47 crc kubenswrapper[4742]: E1004 03:13:47.065292 4742 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 04 03:13:47 crc kubenswrapper[4742]: E1004 03:13:47.065449 4742 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4ssgl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-46h54_openshift-marketplace(7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 03:13:47 crc kubenswrapper[4742]: E1004 03:13:47.066831 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-46h54" podUID="7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a" Oct 04 03:13:47 crc kubenswrapper[4742]: I1004 03:13:47.369017 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-cd46x"] Oct 04 03:13:47 crc kubenswrapper[4742]: W1004 03:13:47.392413 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0302950c_deb3_420a_b798_c99459e62b3f.slice/crio-b4cef373fc1fc79c1c7a5b1e3f068219eab9d76d97c610a7ee2ffb798d0d8970 WatchSource:0}: Error finding container b4cef373fc1fc79c1c7a5b1e3f068219eab9d76d97c610a7ee2ffb798d0d8970: Status 404 returned error can't find the container with id b4cef373fc1fc79c1c7a5b1e3f068219eab9d76d97c610a7ee2ffb798d0d8970 Oct 04 03:13:47 crc kubenswrapper[4742]: I1004 03:13:47.974062 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-cd46x" event={"ID":"0302950c-deb3-420a-b798-c99459e62b3f","Type":"ContainerStarted","Data":"73a62d250bdde0dbb1599f1e54360d780051916c81ab816495a85159fc2f753b"} Oct 04 03:13:47 crc kubenswrapper[4742]: I1004 03:13:47.974120 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-cd46x" event={"ID":"0302950c-deb3-420a-b798-c99459e62b3f","Type":"ContainerStarted","Data":"cc77abacfbd9976c1bc82595c281b3d584d97e4365d99202204cca2171793113"} Oct 04 03:13:47 crc kubenswrapper[4742]: I1004 03:13:47.974133 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-cd46x" event={"ID":"0302950c-deb3-420a-b798-c99459e62b3f","Type":"ContainerStarted","Data":"b4cef373fc1fc79c1c7a5b1e3f068219eab9d76d97c610a7ee2ffb798d0d8970"} Oct 04 03:13:47 crc kubenswrapper[4742]: E1004 03:13:47.977687 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-lb54b" podUID="c3fa495f-e19d-4658-a335-1e49cb14a721" Oct 04 03:13:47 crc kubenswrapper[4742]: E1004 03:13:47.977845 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-46h54" podUID="7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a" Oct 04 03:13:48 crc kubenswrapper[4742]: I1004 03:13:48.028466 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-cd46x" podStartSLOduration=171.028430256 podStartE2EDuration="2m51.028430256s" podCreationTimestamp="2025-10-04 03:10:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:13:48.025905158 +0000 UTC m=+193.958242334" watchObservedRunningTime="2025-10-04 03:13:48.028430256 +0000 UTC m=+193.960767422" Oct 04 03:13:57 crc kubenswrapper[4742]: I1004 03:13:57.022495 4742 generic.go:334] "Generic (PLEG): container finished" podID="2788ab92-5c72-4f27-888e-ef695e4cf6cb" containerID="75886b99c603a77979482570e4ab012ad56aca7cecc570aa77442da750ecef02" exitCode=0 Oct 04 03:13:57 crc kubenswrapper[4742]: I1004 03:13:57.022583 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-trfnh" event={"ID":"2788ab92-5c72-4f27-888e-ef695e4cf6cb","Type":"ContainerDied","Data":"75886b99c603a77979482570e4ab012ad56aca7cecc570aa77442da750ecef02"} Oct 04 03:13:58 crc kubenswrapper[4742]: I1004 03:13:58.031534 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-trfnh" event={"ID":"2788ab92-5c72-4f27-888e-ef695e4cf6cb","Type":"ContainerStarted","Data":"abaa6ec10825e7beaa1a4d045a9d20823ba519919ca5549d131d09c1fb50dee6"} Oct 04 03:13:58 crc kubenswrapper[4742]: I1004 03:13:58.032936 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72m4v" event={"ID":"0fbf09fa-a448-4ac9-8550-e9c719ec22d2","Type":"ContainerStarted","Data":"9a702557a4efce7ef8c3bc117abcd81735f9199e29b5e68332184c6dec7346fa"} Oct 04 03:13:58 crc kubenswrapper[4742]: I1004 03:13:58.054261 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-trfnh" podStartSLOduration=2.996016439 podStartE2EDuration="56.054227855s" podCreationTimestamp="2025-10-04 03:13:02 +0000 UTC" firstStartedPulling="2025-10-04 03:13:04.420785062 +0000 UTC m=+150.353122238" lastFinishedPulling="2025-10-04 03:13:57.478996468 +0000 UTC m=+203.411333654" observedRunningTime="2025-10-04 03:13:58.050478267 +0000 UTC m=+203.982815463" watchObservedRunningTime="2025-10-04 03:13:58.054227855 +0000 UTC m=+203.986565031" Oct 04 03:13:59 crc kubenswrapper[4742]: I1004 03:13:59.041373 4742 generic.go:334] "Generic (PLEG): container finished" podID="9afa0f84-44ab-47b1-986f-b545fddb3824" containerID="99446c8d7779a149533093ddf41040a8ca6ff3b4cf8303d6258ffe5e427ffccf" exitCode=0 Oct 04 03:13:59 crc kubenswrapper[4742]: I1004 03:13:59.042282 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j5nft" event={"ID":"9afa0f84-44ab-47b1-986f-b545fddb3824","Type":"ContainerDied","Data":"99446c8d7779a149533093ddf41040a8ca6ff3b4cf8303d6258ffe5e427ffccf"} Oct 04 03:13:59 crc kubenswrapper[4742]: I1004 03:13:59.059172 4742 generic.go:334] "Generic (PLEG): container finished" podID="0fbf09fa-a448-4ac9-8550-e9c719ec22d2" containerID="9a702557a4efce7ef8c3bc117abcd81735f9199e29b5e68332184c6dec7346fa" exitCode=0 Oct 04 03:13:59 crc kubenswrapper[4742]: I1004 03:13:59.059295 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72m4v" event={"ID":"0fbf09fa-a448-4ac9-8550-e9c719ec22d2","Type":"ContainerDied","Data":"9a702557a4efce7ef8c3bc117abcd81735f9199e29b5e68332184c6dec7346fa"} Oct 04 03:14:01 crc kubenswrapper[4742]: I1004 03:14:01.076220 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j5nft" event={"ID":"9afa0f84-44ab-47b1-986f-b545fddb3824","Type":"ContainerStarted","Data":"8d07d9eabd25cc67c9c28b4cd959959e911800dfb2a050e250c3594f219e82c6"} Oct 04 03:14:01 crc kubenswrapper[4742]: I1004 03:14:01.081344 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72m4v" event={"ID":"0fbf09fa-a448-4ac9-8550-e9c719ec22d2","Type":"ContainerStarted","Data":"30d58be6a89589956fd534cc1fa302bb12c3903703ddb1231753a362941db838"} Oct 04 03:14:01 crc kubenswrapper[4742]: I1004 03:14:01.085320 4742 generic.go:334] "Generic (PLEG): container finished" podID="3848a0f4-0c53-4a32-ad20-d4bce94d9a95" containerID="52f5eba945f2babfdf82d376ea6306dadcaeeb5263443e1e6afcfa6557f995c1" exitCode=0 Oct 04 03:14:01 crc kubenswrapper[4742]: I1004 03:14:01.085355 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rd5tv" event={"ID":"3848a0f4-0c53-4a32-ad20-d4bce94d9a95","Type":"ContainerDied","Data":"52f5eba945f2babfdf82d376ea6306dadcaeeb5263443e1e6afcfa6557f995c1"} Oct 04 03:14:01 crc kubenswrapper[4742]: I1004 03:14:01.105787 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-j5nft" podStartSLOduration=3.875123169 podStartE2EDuration="58.105753161s" podCreationTimestamp="2025-10-04 03:13:03 +0000 UTC" firstStartedPulling="2025-10-04 03:13:05.577954189 +0000 UTC m=+151.510291365" lastFinishedPulling="2025-10-04 03:13:59.808584181 +0000 UTC m=+205.740921357" observedRunningTime="2025-10-04 03:14:01.103967304 +0000 UTC m=+207.036304490" watchObservedRunningTime="2025-10-04 03:14:01.105753161 +0000 UTC m=+207.038090337" Oct 04 03:14:01 crc kubenswrapper[4742]: I1004 03:14:01.877117 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-72m4v" podStartSLOduration=3.473980719 podStartE2EDuration="58.877086255s" podCreationTimestamp="2025-10-04 03:13:03 +0000 UTC" firstStartedPulling="2025-10-04 03:13:04.480093561 +0000 UTC m=+150.412430737" lastFinishedPulling="2025-10-04 03:13:59.883199097 +0000 UTC m=+205.815536273" observedRunningTime="2025-10-04 03:14:01.153662433 +0000 UTC m=+207.085999609" watchObservedRunningTime="2025-10-04 03:14:01.877086255 +0000 UTC m=+207.809423431" Oct 04 03:14:03 crc kubenswrapper[4742]: I1004 03:14:03.169332 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-trfnh" Oct 04 03:14:03 crc kubenswrapper[4742]: I1004 03:14:03.169796 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-trfnh" Oct 04 03:14:03 crc kubenswrapper[4742]: I1004 03:14:03.585177 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-72m4v" Oct 04 03:14:03 crc kubenswrapper[4742]: I1004 03:14:03.585912 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-72m4v" Oct 04 03:14:03 crc kubenswrapper[4742]: I1004 03:14:03.669057 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-trfnh" Oct 04 03:14:03 crc kubenswrapper[4742]: I1004 03:14:03.669197 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-72m4v" Oct 04 03:14:03 crc kubenswrapper[4742]: I1004 03:14:03.911645 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-j5nft" Oct 04 03:14:03 crc kubenswrapper[4742]: I1004 03:14:03.911782 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-j5nft" Oct 04 03:14:03 crc kubenswrapper[4742]: I1004 03:14:03.955315 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-j5nft" Oct 04 03:14:04 crc kubenswrapper[4742]: I1004 03:14:04.145544 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-trfnh" Oct 04 03:14:05 crc kubenswrapper[4742]: I1004 03:14:05.152743 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-j5nft" Oct 04 03:14:05 crc kubenswrapper[4742]: I1004 03:14:05.153814 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-72m4v" Oct 04 03:14:07 crc kubenswrapper[4742]: I1004 03:14:07.101780 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j5nft"] Oct 04 03:14:07 crc kubenswrapper[4742]: I1004 03:14:07.116612 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-j5nft" podUID="9afa0f84-44ab-47b1-986f-b545fddb3824" containerName="registry-server" containerID="cri-o://8d07d9eabd25cc67c9c28b4cd959959e911800dfb2a050e250c3594f219e82c6" gracePeriod=2 Oct 04 03:14:09 crc kubenswrapper[4742]: I1004 03:14:09.131950 4742 generic.go:334] "Generic (PLEG): container finished" podID="9afa0f84-44ab-47b1-986f-b545fddb3824" containerID="8d07d9eabd25cc67c9c28b4cd959959e911800dfb2a050e250c3594f219e82c6" exitCode=0 Oct 04 03:14:09 crc kubenswrapper[4742]: I1004 03:14:09.132429 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j5nft" event={"ID":"9afa0f84-44ab-47b1-986f-b545fddb3824","Type":"ContainerDied","Data":"8d07d9eabd25cc67c9c28b4cd959959e911800dfb2a050e250c3594f219e82c6"} Oct 04 03:14:09 crc kubenswrapper[4742]: I1004 03:14:09.300025 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-72m4v"] Oct 04 03:14:09 crc kubenswrapper[4742]: I1004 03:14:09.300659 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-72m4v" podUID="0fbf09fa-a448-4ac9-8550-e9c719ec22d2" containerName="registry-server" containerID="cri-o://30d58be6a89589956fd534cc1fa302bb12c3903703ddb1231753a362941db838" gracePeriod=2 Oct 04 03:14:10 crc kubenswrapper[4742]: I1004 03:14:10.664316 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j5nft" Oct 04 03:14:10 crc kubenswrapper[4742]: I1004 03:14:10.709787 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9afa0f84-44ab-47b1-986f-b545fddb3824-catalog-content\") pod \"9afa0f84-44ab-47b1-986f-b545fddb3824\" (UID: \"9afa0f84-44ab-47b1-986f-b545fddb3824\") " Oct 04 03:14:10 crc kubenswrapper[4742]: I1004 03:14:10.709851 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7l4mf\" (UniqueName: \"kubernetes.io/projected/9afa0f84-44ab-47b1-986f-b545fddb3824-kube-api-access-7l4mf\") pod \"9afa0f84-44ab-47b1-986f-b545fddb3824\" (UID: \"9afa0f84-44ab-47b1-986f-b545fddb3824\") " Oct 04 03:14:10 crc kubenswrapper[4742]: I1004 03:14:10.709919 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9afa0f84-44ab-47b1-986f-b545fddb3824-utilities\") pod \"9afa0f84-44ab-47b1-986f-b545fddb3824\" (UID: \"9afa0f84-44ab-47b1-986f-b545fddb3824\") " Oct 04 03:14:10 crc kubenswrapper[4742]: I1004 03:14:10.710850 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9afa0f84-44ab-47b1-986f-b545fddb3824-utilities" (OuterVolumeSpecName: "utilities") pod "9afa0f84-44ab-47b1-986f-b545fddb3824" (UID: "9afa0f84-44ab-47b1-986f-b545fddb3824"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:14:10 crc kubenswrapper[4742]: I1004 03:14:10.716090 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9afa0f84-44ab-47b1-986f-b545fddb3824-kube-api-access-7l4mf" (OuterVolumeSpecName: "kube-api-access-7l4mf") pod "9afa0f84-44ab-47b1-986f-b545fddb3824" (UID: "9afa0f84-44ab-47b1-986f-b545fddb3824"). InnerVolumeSpecName "kube-api-access-7l4mf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:14:10 crc kubenswrapper[4742]: I1004 03:14:10.811469 4742 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9afa0f84-44ab-47b1-986f-b545fddb3824-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:10 crc kubenswrapper[4742]: I1004 03:14:10.811503 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7l4mf\" (UniqueName: \"kubernetes.io/projected/9afa0f84-44ab-47b1-986f-b545fddb3824-kube-api-access-7l4mf\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:10 crc kubenswrapper[4742]: I1004 03:14:10.933048 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-72m4v" Oct 04 03:14:11 crc kubenswrapper[4742]: I1004 03:14:11.014365 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5f6j\" (UniqueName: \"kubernetes.io/projected/0fbf09fa-a448-4ac9-8550-e9c719ec22d2-kube-api-access-x5f6j\") pod \"0fbf09fa-a448-4ac9-8550-e9c719ec22d2\" (UID: \"0fbf09fa-a448-4ac9-8550-e9c719ec22d2\") " Oct 04 03:14:11 crc kubenswrapper[4742]: I1004 03:14:11.014499 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fbf09fa-a448-4ac9-8550-e9c719ec22d2-catalog-content\") pod \"0fbf09fa-a448-4ac9-8550-e9c719ec22d2\" (UID: \"0fbf09fa-a448-4ac9-8550-e9c719ec22d2\") " Oct 04 03:14:11 crc kubenswrapper[4742]: I1004 03:14:11.014552 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fbf09fa-a448-4ac9-8550-e9c719ec22d2-utilities\") pod \"0fbf09fa-a448-4ac9-8550-e9c719ec22d2\" (UID: \"0fbf09fa-a448-4ac9-8550-e9c719ec22d2\") " Oct 04 03:14:11 crc kubenswrapper[4742]: I1004 03:14:11.015342 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fbf09fa-a448-4ac9-8550-e9c719ec22d2-utilities" (OuterVolumeSpecName: "utilities") pod "0fbf09fa-a448-4ac9-8550-e9c719ec22d2" (UID: "0fbf09fa-a448-4ac9-8550-e9c719ec22d2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:14:11 crc kubenswrapper[4742]: I1004 03:14:11.017540 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fbf09fa-a448-4ac9-8550-e9c719ec22d2-kube-api-access-x5f6j" (OuterVolumeSpecName: "kube-api-access-x5f6j") pod "0fbf09fa-a448-4ac9-8550-e9c719ec22d2" (UID: "0fbf09fa-a448-4ac9-8550-e9c719ec22d2"). InnerVolumeSpecName "kube-api-access-x5f6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:14:11 crc kubenswrapper[4742]: I1004 03:14:11.116535 4742 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fbf09fa-a448-4ac9-8550-e9c719ec22d2-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:11 crc kubenswrapper[4742]: I1004 03:14:11.116599 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5f6j\" (UniqueName: \"kubernetes.io/projected/0fbf09fa-a448-4ac9-8550-e9c719ec22d2-kube-api-access-x5f6j\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:11 crc kubenswrapper[4742]: I1004 03:14:11.143289 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j5nft" event={"ID":"9afa0f84-44ab-47b1-986f-b545fddb3824","Type":"ContainerDied","Data":"eaee1d4a0823e5f19f946a1fab6da4d3821df5e36fd1854e8011973683532e18"} Oct 04 03:14:11 crc kubenswrapper[4742]: I1004 03:14:11.143338 4742 scope.go:117] "RemoveContainer" containerID="8d07d9eabd25cc67c9c28b4cd959959e911800dfb2a050e250c3594f219e82c6" Oct 04 03:14:11 crc kubenswrapper[4742]: I1004 03:14:11.143436 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j5nft" Oct 04 03:14:11 crc kubenswrapper[4742]: I1004 03:14:11.148059 4742 generic.go:334] "Generic (PLEG): container finished" podID="0fbf09fa-a448-4ac9-8550-e9c719ec22d2" containerID="30d58be6a89589956fd534cc1fa302bb12c3903703ddb1231753a362941db838" exitCode=0 Oct 04 03:14:11 crc kubenswrapper[4742]: I1004 03:14:11.148159 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-72m4v" Oct 04 03:14:11 crc kubenswrapper[4742]: I1004 03:14:11.148146 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72m4v" event={"ID":"0fbf09fa-a448-4ac9-8550-e9c719ec22d2","Type":"ContainerDied","Data":"30d58be6a89589956fd534cc1fa302bb12c3903703ddb1231753a362941db838"} Oct 04 03:14:11 crc kubenswrapper[4742]: I1004 03:14:11.148355 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72m4v" event={"ID":"0fbf09fa-a448-4ac9-8550-e9c719ec22d2","Type":"ContainerDied","Data":"d72d75824473a8968c7b350ddf34e4f23193dad3c6bbd939045abbb9ad923fdd"} Oct 04 03:14:11 crc kubenswrapper[4742]: I1004 03:14:11.184218 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9afa0f84-44ab-47b1-986f-b545fddb3824-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9afa0f84-44ab-47b1-986f-b545fddb3824" (UID: "9afa0f84-44ab-47b1-986f-b545fddb3824"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:14:11 crc kubenswrapper[4742]: I1004 03:14:11.218176 4742 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9afa0f84-44ab-47b1-986f-b545fddb3824-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:11 crc kubenswrapper[4742]: I1004 03:14:11.458937 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fbf09fa-a448-4ac9-8550-e9c719ec22d2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0fbf09fa-a448-4ac9-8550-e9c719ec22d2" (UID: "0fbf09fa-a448-4ac9-8550-e9c719ec22d2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:14:11 crc kubenswrapper[4742]: I1004 03:14:11.470296 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j5nft"] Oct 04 03:14:11 crc kubenswrapper[4742]: I1004 03:14:11.473614 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-j5nft"] Oct 04 03:14:11 crc kubenswrapper[4742]: I1004 03:14:11.522992 4742 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fbf09fa-a448-4ac9-8550-e9c719ec22d2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:11 crc kubenswrapper[4742]: I1004 03:14:11.659556 4742 scope.go:117] "RemoveContainer" containerID="99446c8d7779a149533093ddf41040a8ca6ff3b4cf8303d6258ffe5e427ffccf" Oct 04 03:14:11 crc kubenswrapper[4742]: I1004 03:14:11.676714 4742 scope.go:117] "RemoveContainer" containerID="4a2a982cee2ee0cf2fb0c8e8613b45a4b460695fc00f91d5cd3ae5a35f47305c" Oct 04 03:14:11 crc kubenswrapper[4742]: I1004 03:14:11.752607 4742 scope.go:117] "RemoveContainer" containerID="30d58be6a89589956fd534cc1fa302bb12c3903703ddb1231753a362941db838" Oct 04 03:14:11 crc kubenswrapper[4742]: I1004 03:14:11.830532 4742 scope.go:117] "RemoveContainer" containerID="9a702557a4efce7ef8c3bc117abcd81735f9199e29b5e68332184c6dec7346fa" Oct 04 03:14:11 crc kubenswrapper[4742]: I1004 03:14:11.866049 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-72m4v"] Oct 04 03:14:11 crc kubenswrapper[4742]: I1004 03:14:11.871243 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-72m4v"] Oct 04 03:14:11 crc kubenswrapper[4742]: I1004 03:14:11.880354 4742 scope.go:117] "RemoveContainer" containerID="26381f12e5108833d45d4c0aa97495c62d1f48d8f66c758baea0bf01127b3c03" Oct 04 03:14:11 crc kubenswrapper[4742]: I1004 03:14:11.909592 4742 scope.go:117] "RemoveContainer" containerID="30d58be6a89589956fd534cc1fa302bb12c3903703ddb1231753a362941db838" Oct 04 03:14:11 crc kubenswrapper[4742]: E1004 03:14:11.910225 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30d58be6a89589956fd534cc1fa302bb12c3903703ddb1231753a362941db838\": container with ID starting with 30d58be6a89589956fd534cc1fa302bb12c3903703ddb1231753a362941db838 not found: ID does not exist" containerID="30d58be6a89589956fd534cc1fa302bb12c3903703ddb1231753a362941db838" Oct 04 03:14:11 crc kubenswrapper[4742]: I1004 03:14:11.910381 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30d58be6a89589956fd534cc1fa302bb12c3903703ddb1231753a362941db838"} err="failed to get container status \"30d58be6a89589956fd534cc1fa302bb12c3903703ddb1231753a362941db838\": rpc error: code = NotFound desc = could not find container \"30d58be6a89589956fd534cc1fa302bb12c3903703ddb1231753a362941db838\": container with ID starting with 30d58be6a89589956fd534cc1fa302bb12c3903703ddb1231753a362941db838 not found: ID does not exist" Oct 04 03:14:11 crc kubenswrapper[4742]: I1004 03:14:11.910573 4742 scope.go:117] "RemoveContainer" containerID="9a702557a4efce7ef8c3bc117abcd81735f9199e29b5e68332184c6dec7346fa" Oct 04 03:14:11 crc kubenswrapper[4742]: E1004 03:14:11.911118 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a702557a4efce7ef8c3bc117abcd81735f9199e29b5e68332184c6dec7346fa\": container with ID starting with 9a702557a4efce7ef8c3bc117abcd81735f9199e29b5e68332184c6dec7346fa not found: ID does not exist" containerID="9a702557a4efce7ef8c3bc117abcd81735f9199e29b5e68332184c6dec7346fa" Oct 04 03:14:11 crc kubenswrapper[4742]: I1004 03:14:11.911200 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a702557a4efce7ef8c3bc117abcd81735f9199e29b5e68332184c6dec7346fa"} err="failed to get container status \"9a702557a4efce7ef8c3bc117abcd81735f9199e29b5e68332184c6dec7346fa\": rpc error: code = NotFound desc = could not find container \"9a702557a4efce7ef8c3bc117abcd81735f9199e29b5e68332184c6dec7346fa\": container with ID starting with 9a702557a4efce7ef8c3bc117abcd81735f9199e29b5e68332184c6dec7346fa not found: ID does not exist" Oct 04 03:14:11 crc kubenswrapper[4742]: I1004 03:14:11.911265 4742 scope.go:117] "RemoveContainer" containerID="26381f12e5108833d45d4c0aa97495c62d1f48d8f66c758baea0bf01127b3c03" Oct 04 03:14:11 crc kubenswrapper[4742]: E1004 03:14:11.915402 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26381f12e5108833d45d4c0aa97495c62d1f48d8f66c758baea0bf01127b3c03\": container with ID starting with 26381f12e5108833d45d4c0aa97495c62d1f48d8f66c758baea0bf01127b3c03 not found: ID does not exist" containerID="26381f12e5108833d45d4c0aa97495c62d1f48d8f66c758baea0bf01127b3c03" Oct 04 03:14:11 crc kubenswrapper[4742]: I1004 03:14:11.915451 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26381f12e5108833d45d4c0aa97495c62d1f48d8f66c758baea0bf01127b3c03"} err="failed to get container status \"26381f12e5108833d45d4c0aa97495c62d1f48d8f66c758baea0bf01127b3c03\": rpc error: code = NotFound desc = could not find container \"26381f12e5108833d45d4c0aa97495c62d1f48d8f66c758baea0bf01127b3c03\": container with ID starting with 26381f12e5108833d45d4c0aa97495c62d1f48d8f66c758baea0bf01127b3c03 not found: ID does not exist" Oct 04 03:14:12 crc kubenswrapper[4742]: I1004 03:14:12.867669 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fbf09fa-a448-4ac9-8550-e9c719ec22d2" path="/var/lib/kubelet/pods/0fbf09fa-a448-4ac9-8550-e9c719ec22d2/volumes" Oct 04 03:14:12 crc kubenswrapper[4742]: I1004 03:14:12.868656 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9afa0f84-44ab-47b1-986f-b545fddb3824" path="/var/lib/kubelet/pods/9afa0f84-44ab-47b1-986f-b545fddb3824/volumes" Oct 04 03:14:13 crc kubenswrapper[4742]: I1004 03:14:13.159767 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rd5tv" event={"ID":"3848a0f4-0c53-4a32-ad20-d4bce94d9a95","Type":"ContainerStarted","Data":"d7160de9da2516337c804dbbf975242d9409b13737e5d8c90eacbf552c9b5c85"} Oct 04 03:14:13 crc kubenswrapper[4742]: I1004 03:14:13.163321 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-67ztt" event={"ID":"c210acab-8945-4f9d-a8fe-47ae7d78a064","Type":"ContainerStarted","Data":"a0b830382262c3de530a309506e7061b48f3467ae89d38d2e2ffca8b801be0ea"} Oct 04 03:14:13 crc kubenswrapper[4742]: I1004 03:14:13.164768 4742 generic.go:334] "Generic (PLEG): container finished" podID="c3fa495f-e19d-4658-a335-1e49cb14a721" containerID="5dde8a9c7c483d0a345f79d13dc330492b90fd476cf879785279928d4e40982a" exitCode=0 Oct 04 03:14:13 crc kubenswrapper[4742]: I1004 03:14:13.164819 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lb54b" event={"ID":"c3fa495f-e19d-4658-a335-1e49cb14a721","Type":"ContainerDied","Data":"5dde8a9c7c483d0a345f79d13dc330492b90fd476cf879785279928d4e40982a"} Oct 04 03:14:13 crc kubenswrapper[4742]: I1004 03:14:13.167469 4742 generic.go:334] "Generic (PLEG): container finished" podID="7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a" containerID="64b36a7156b087048338c525070e653d57bf666a0aa51d83ba8bcda021046ea6" exitCode=0 Oct 04 03:14:13 crc kubenswrapper[4742]: I1004 03:14:13.167520 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-46h54" event={"ID":"7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a","Type":"ContainerDied","Data":"64b36a7156b087048338c525070e653d57bf666a0aa51d83ba8bcda021046ea6"} Oct 04 03:14:13 crc kubenswrapper[4742]: I1004 03:14:13.181744 4742 generic.go:334] "Generic (PLEG): container finished" podID="f3b8a6f1-47db-4541-af94-5519b15d1407" containerID="1e188d2c4d79496549160f22be90e4a093df16f4663a5910ba90d607e2ea6f86" exitCode=0 Oct 04 03:14:13 crc kubenswrapper[4742]: I1004 03:14:13.181797 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b5xzp" event={"ID":"f3b8a6f1-47db-4541-af94-5519b15d1407","Type":"ContainerDied","Data":"1e188d2c4d79496549160f22be90e4a093df16f4663a5910ba90d607e2ea6f86"} Oct 04 03:14:13 crc kubenswrapper[4742]: I1004 03:14:13.183700 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rd5tv" podStartSLOduration=10.725372310000001 podStartE2EDuration="1m7.183688467s" podCreationTimestamp="2025-10-04 03:13:06 +0000 UTC" firstStartedPulling="2025-10-04 03:13:09.693241758 +0000 UTC m=+155.625578934" lastFinishedPulling="2025-10-04 03:14:06.151557915 +0000 UTC m=+212.083895091" observedRunningTime="2025-10-04 03:14:13.180337219 +0000 UTC m=+219.112674395" watchObservedRunningTime="2025-10-04 03:14:13.183688467 +0000 UTC m=+219.116025633" Oct 04 03:14:14 crc kubenswrapper[4742]: I1004 03:14:14.197539 4742 generic.go:334] "Generic (PLEG): container finished" podID="c210acab-8945-4f9d-a8fe-47ae7d78a064" containerID="a0b830382262c3de530a309506e7061b48f3467ae89d38d2e2ffca8b801be0ea" exitCode=0 Oct 04 03:14:14 crc kubenswrapper[4742]: I1004 03:14:14.197603 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-67ztt" event={"ID":"c210acab-8945-4f9d-a8fe-47ae7d78a064","Type":"ContainerDied","Data":"a0b830382262c3de530a309506e7061b48f3467ae89d38d2e2ffca8b801be0ea"} Oct 04 03:14:14 crc kubenswrapper[4742]: I1004 03:14:14.202523 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lb54b" event={"ID":"c3fa495f-e19d-4658-a335-1e49cb14a721","Type":"ContainerStarted","Data":"4acb3da1f52b48c905c1821afc3f6102d530d39dfc73cc8b87ba3ef724b85dff"} Oct 04 03:14:14 crc kubenswrapper[4742]: I1004 03:14:14.206188 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-46h54" event={"ID":"7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a","Type":"ContainerStarted","Data":"f10737f977cd408e311c63ac5a261074c37e42f4a159826d0cfb36e8eaeff284"} Oct 04 03:14:14 crc kubenswrapper[4742]: I1004 03:14:14.208206 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b5xzp" event={"ID":"f3b8a6f1-47db-4541-af94-5519b15d1407","Type":"ContainerStarted","Data":"052b37f81317c68a03ea64f3337d7574c49e74724e2c21557dabc01db252e9aa"} Oct 04 03:14:14 crc kubenswrapper[4742]: I1004 03:14:14.250805 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lb54b" podStartSLOduration=3.304133904 podStartE2EDuration="1m10.250786644s" podCreationTimestamp="2025-10-04 03:13:04 +0000 UTC" firstStartedPulling="2025-10-04 03:13:06.615937511 +0000 UTC m=+152.548274687" lastFinishedPulling="2025-10-04 03:14:13.562590261 +0000 UTC m=+219.494927427" observedRunningTime="2025-10-04 03:14:14.247109007 +0000 UTC m=+220.179446193" watchObservedRunningTime="2025-10-04 03:14:14.250786644 +0000 UTC m=+220.183123820" Oct 04 03:14:14 crc kubenswrapper[4742]: I1004 03:14:14.272460 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-b5xzp" podStartSLOduration=2.145464835 podStartE2EDuration="1m11.272446204s" podCreationTimestamp="2025-10-04 03:13:03 +0000 UTC" firstStartedPulling="2025-10-04 03:13:04.487553741 +0000 UTC m=+150.419890917" lastFinishedPulling="2025-10-04 03:14:13.61453511 +0000 UTC m=+219.546872286" observedRunningTime="2025-10-04 03:14:14.269458116 +0000 UTC m=+220.201795302" watchObservedRunningTime="2025-10-04 03:14:14.272446204 +0000 UTC m=+220.204783380" Oct 04 03:14:14 crc kubenswrapper[4742]: I1004 03:14:14.872523 4742 patch_prober.go:28] interesting pod/machine-config-daemon-qdt7v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:14:14 crc kubenswrapper[4742]: I1004 03:14:14.872604 4742 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:14:14 crc kubenswrapper[4742]: I1004 03:14:14.872661 4742 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" Oct 04 03:14:14 crc kubenswrapper[4742]: I1004 03:14:14.873410 4742 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde"} pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 03:14:14 crc kubenswrapper[4742]: I1004 03:14:14.873484 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerName="machine-config-daemon" containerID="cri-o://dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde" gracePeriod=600 Oct 04 03:14:15 crc kubenswrapper[4742]: I1004 03:14:15.182957 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lb54b" Oct 04 03:14:15 crc kubenswrapper[4742]: I1004 03:14:15.183241 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lb54b" Oct 04 03:14:15 crc kubenswrapper[4742]: I1004 03:14:15.219138 4742 generic.go:334] "Generic (PLEG): container finished" podID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerID="dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde" exitCode=0 Oct 04 03:14:15 crc kubenswrapper[4742]: I1004 03:14:15.219191 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" event={"ID":"989ea247-7ac9-4ade-8dcb-2632fef4f83e","Type":"ContainerDied","Data":"dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde"} Oct 04 03:14:15 crc kubenswrapper[4742]: I1004 03:14:15.219217 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" event={"ID":"989ea247-7ac9-4ade-8dcb-2632fef4f83e","Type":"ContainerStarted","Data":"265d94a26e57f0eae37e63999ff3e9ea88f239556449abe4f5a9ca7688a659d5"} Oct 04 03:14:15 crc kubenswrapper[4742]: I1004 03:14:15.222457 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-67ztt" event={"ID":"c210acab-8945-4f9d-a8fe-47ae7d78a064","Type":"ContainerStarted","Data":"dafdecd26505ab05deade27027f7e144160fa80a5642e21b0fcc722372082dcd"} Oct 04 03:14:15 crc kubenswrapper[4742]: I1004 03:14:15.235599 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lb54b" Oct 04 03:14:15 crc kubenswrapper[4742]: I1004 03:14:15.245388 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-46h54" podStartSLOduration=3.232425484 podStartE2EDuration="1m10.2453574s" podCreationTimestamp="2025-10-04 03:13:05 +0000 UTC" firstStartedPulling="2025-10-04 03:13:06.616012623 +0000 UTC m=+152.548349799" lastFinishedPulling="2025-10-04 03:14:13.628944539 +0000 UTC m=+219.561281715" observedRunningTime="2025-10-04 03:14:14.288549229 +0000 UTC m=+220.220886415" watchObservedRunningTime="2025-10-04 03:14:15.2453574 +0000 UTC m=+221.177694576" Oct 04 03:14:15 crc kubenswrapper[4742]: I1004 03:14:15.581083 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-46h54" Oct 04 03:14:15 crc kubenswrapper[4742]: I1004 03:14:15.581704 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-46h54" Oct 04 03:14:15 crc kubenswrapper[4742]: I1004 03:14:15.624975 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-46h54" Oct 04 03:14:15 crc kubenswrapper[4742]: I1004 03:14:15.648876 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-67ztt" podStartSLOduration=2.6420496079999998 podStartE2EDuration="1m9.648853112s" podCreationTimestamp="2025-10-04 03:13:06 +0000 UTC" firstStartedPulling="2025-10-04 03:13:07.647674407 +0000 UTC m=+153.580011583" lastFinishedPulling="2025-10-04 03:14:14.654477911 +0000 UTC m=+220.586815087" observedRunningTime="2025-10-04 03:14:15.294439843 +0000 UTC m=+221.226777029" watchObservedRunningTime="2025-10-04 03:14:15.648853112 +0000 UTC m=+221.581190288" Oct 04 03:14:16 crc kubenswrapper[4742]: I1004 03:14:16.587108 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-67ztt" Oct 04 03:14:16 crc kubenswrapper[4742]: I1004 03:14:16.587156 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-67ztt" Oct 04 03:14:17 crc kubenswrapper[4742]: I1004 03:14:17.022874 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rd5tv" Oct 04 03:14:17 crc kubenswrapper[4742]: I1004 03:14:17.023279 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rd5tv" Oct 04 03:14:17 crc kubenswrapper[4742]: I1004 03:14:17.624191 4742 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-67ztt" podUID="c210acab-8945-4f9d-a8fe-47ae7d78a064" containerName="registry-server" probeResult="failure" output=< Oct 04 03:14:17 crc kubenswrapper[4742]: timeout: failed to connect service ":50051" within 1s Oct 04 03:14:17 crc kubenswrapper[4742]: > Oct 04 03:14:18 crc kubenswrapper[4742]: I1004 03:14:18.063299 4742 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rd5tv" podUID="3848a0f4-0c53-4a32-ad20-d4bce94d9a95" containerName="registry-server" probeResult="failure" output=< Oct 04 03:14:18 crc kubenswrapper[4742]: timeout: failed to connect service ":50051" within 1s Oct 04 03:14:18 crc kubenswrapper[4742]: > Oct 04 03:14:23 crc kubenswrapper[4742]: I1004 03:14:23.369408 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-b5xzp" Oct 04 03:14:23 crc kubenswrapper[4742]: I1004 03:14:23.370078 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-b5xzp" Oct 04 03:14:23 crc kubenswrapper[4742]: I1004 03:14:23.414805 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-b5xzp" Oct 04 03:14:24 crc kubenswrapper[4742]: I1004 03:14:24.316570 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-b5xzp" Oct 04 03:14:25 crc kubenswrapper[4742]: I1004 03:14:25.029932 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mspvb"] Oct 04 03:14:25 crc kubenswrapper[4742]: I1004 03:14:25.219351 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lb54b" Oct 04 03:14:25 crc kubenswrapper[4742]: I1004 03:14:25.618354 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-46h54" Oct 04 03:14:26 crc kubenswrapper[4742]: I1004 03:14:26.498911 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-46h54"] Oct 04 03:14:26 crc kubenswrapper[4742]: I1004 03:14:26.499135 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-46h54" podUID="7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a" containerName="registry-server" containerID="cri-o://f10737f977cd408e311c63ac5a261074c37e42f4a159826d0cfb36e8eaeff284" gracePeriod=2 Oct 04 03:14:26 crc kubenswrapper[4742]: I1004 03:14:26.624983 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-67ztt" Oct 04 03:14:26 crc kubenswrapper[4742]: I1004 03:14:26.670264 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-67ztt" Oct 04 03:14:26 crc kubenswrapper[4742]: I1004 03:14:26.833557 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-46h54" Oct 04 03:14:26 crc kubenswrapper[4742]: I1004 03:14:26.914709 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a-utilities\") pod \"7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a\" (UID: \"7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a\") " Oct 04 03:14:26 crc kubenswrapper[4742]: I1004 03:14:26.914771 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a-catalog-content\") pod \"7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a\" (UID: \"7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a\") " Oct 04 03:14:26 crc kubenswrapper[4742]: I1004 03:14:26.914911 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ssgl\" (UniqueName: \"kubernetes.io/projected/7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a-kube-api-access-4ssgl\") pod \"7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a\" (UID: \"7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a\") " Oct 04 03:14:26 crc kubenswrapper[4742]: I1004 03:14:26.915548 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a-utilities" (OuterVolumeSpecName: "utilities") pod "7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a" (UID: "7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:14:26 crc kubenswrapper[4742]: I1004 03:14:26.920280 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a-kube-api-access-4ssgl" (OuterVolumeSpecName: "kube-api-access-4ssgl") pod "7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a" (UID: "7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a"). InnerVolumeSpecName "kube-api-access-4ssgl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:14:26 crc kubenswrapper[4742]: I1004 03:14:26.927420 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a" (UID: "7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:14:27 crc kubenswrapper[4742]: I1004 03:14:27.016519 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ssgl\" (UniqueName: \"kubernetes.io/projected/7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a-kube-api-access-4ssgl\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:27 crc kubenswrapper[4742]: I1004 03:14:27.016553 4742 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:27 crc kubenswrapper[4742]: I1004 03:14:27.016563 4742 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:27 crc kubenswrapper[4742]: I1004 03:14:27.062768 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rd5tv" Oct 04 03:14:27 crc kubenswrapper[4742]: I1004 03:14:27.105261 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rd5tv" Oct 04 03:14:27 crc kubenswrapper[4742]: I1004 03:14:27.290537 4742 generic.go:334] "Generic (PLEG): container finished" podID="7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a" containerID="f10737f977cd408e311c63ac5a261074c37e42f4a159826d0cfb36e8eaeff284" exitCode=0 Oct 04 03:14:27 crc kubenswrapper[4742]: I1004 03:14:27.290625 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-46h54" Oct 04 03:14:27 crc kubenswrapper[4742]: I1004 03:14:27.290619 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-46h54" event={"ID":"7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a","Type":"ContainerDied","Data":"f10737f977cd408e311c63ac5a261074c37e42f4a159826d0cfb36e8eaeff284"} Oct 04 03:14:27 crc kubenswrapper[4742]: I1004 03:14:27.290758 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-46h54" event={"ID":"7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a","Type":"ContainerDied","Data":"835a760f9d53c3a1034dff098d37b9e473eaa73674cf462fdfa40d3db20d0559"} Oct 04 03:14:27 crc kubenswrapper[4742]: I1004 03:14:27.290786 4742 scope.go:117] "RemoveContainer" containerID="f10737f977cd408e311c63ac5a261074c37e42f4a159826d0cfb36e8eaeff284" Oct 04 03:14:27 crc kubenswrapper[4742]: I1004 03:14:27.309707 4742 scope.go:117] "RemoveContainer" containerID="64b36a7156b087048338c525070e653d57bf666a0aa51d83ba8bcda021046ea6" Oct 04 03:14:27 crc kubenswrapper[4742]: I1004 03:14:27.330938 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-46h54"] Oct 04 03:14:27 crc kubenswrapper[4742]: I1004 03:14:27.334966 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-46h54"] Oct 04 03:14:27 crc kubenswrapper[4742]: I1004 03:14:27.359574 4742 scope.go:117] "RemoveContainer" containerID="fc957861f4ebcf5836dbed163010972ff837169e0506a8b5877af8849336c973" Oct 04 03:14:27 crc kubenswrapper[4742]: I1004 03:14:27.381374 4742 scope.go:117] "RemoveContainer" containerID="f10737f977cd408e311c63ac5a261074c37e42f4a159826d0cfb36e8eaeff284" Oct 04 03:14:27 crc kubenswrapper[4742]: E1004 03:14:27.383236 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f10737f977cd408e311c63ac5a261074c37e42f4a159826d0cfb36e8eaeff284\": container with ID starting with f10737f977cd408e311c63ac5a261074c37e42f4a159826d0cfb36e8eaeff284 not found: ID does not exist" containerID="f10737f977cd408e311c63ac5a261074c37e42f4a159826d0cfb36e8eaeff284" Oct 04 03:14:27 crc kubenswrapper[4742]: I1004 03:14:27.383339 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f10737f977cd408e311c63ac5a261074c37e42f4a159826d0cfb36e8eaeff284"} err="failed to get container status \"f10737f977cd408e311c63ac5a261074c37e42f4a159826d0cfb36e8eaeff284\": rpc error: code = NotFound desc = could not find container \"f10737f977cd408e311c63ac5a261074c37e42f4a159826d0cfb36e8eaeff284\": container with ID starting with f10737f977cd408e311c63ac5a261074c37e42f4a159826d0cfb36e8eaeff284 not found: ID does not exist" Oct 04 03:14:27 crc kubenswrapper[4742]: I1004 03:14:27.383375 4742 scope.go:117] "RemoveContainer" containerID="64b36a7156b087048338c525070e653d57bf666a0aa51d83ba8bcda021046ea6" Oct 04 03:14:27 crc kubenswrapper[4742]: E1004 03:14:27.383786 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64b36a7156b087048338c525070e653d57bf666a0aa51d83ba8bcda021046ea6\": container with ID starting with 64b36a7156b087048338c525070e653d57bf666a0aa51d83ba8bcda021046ea6 not found: ID does not exist" containerID="64b36a7156b087048338c525070e653d57bf666a0aa51d83ba8bcda021046ea6" Oct 04 03:14:27 crc kubenswrapper[4742]: I1004 03:14:27.383837 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64b36a7156b087048338c525070e653d57bf666a0aa51d83ba8bcda021046ea6"} err="failed to get container status \"64b36a7156b087048338c525070e653d57bf666a0aa51d83ba8bcda021046ea6\": rpc error: code = NotFound desc = could not find container \"64b36a7156b087048338c525070e653d57bf666a0aa51d83ba8bcda021046ea6\": container with ID starting with 64b36a7156b087048338c525070e653d57bf666a0aa51d83ba8bcda021046ea6 not found: ID does not exist" Oct 04 03:14:27 crc kubenswrapper[4742]: I1004 03:14:27.383877 4742 scope.go:117] "RemoveContainer" containerID="fc957861f4ebcf5836dbed163010972ff837169e0506a8b5877af8849336c973" Oct 04 03:14:27 crc kubenswrapper[4742]: E1004 03:14:27.384341 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc957861f4ebcf5836dbed163010972ff837169e0506a8b5877af8849336c973\": container with ID starting with fc957861f4ebcf5836dbed163010972ff837169e0506a8b5877af8849336c973 not found: ID does not exist" containerID="fc957861f4ebcf5836dbed163010972ff837169e0506a8b5877af8849336c973" Oct 04 03:14:27 crc kubenswrapper[4742]: I1004 03:14:27.384370 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc957861f4ebcf5836dbed163010972ff837169e0506a8b5877af8849336c973"} err="failed to get container status \"fc957861f4ebcf5836dbed163010972ff837169e0506a8b5877af8849336c973\": rpc error: code = NotFound desc = could not find container \"fc957861f4ebcf5836dbed163010972ff837169e0506a8b5877af8849336c973\": container with ID starting with fc957861f4ebcf5836dbed163010972ff837169e0506a8b5877af8849336c973 not found: ID does not exist" Oct 04 03:14:28 crc kubenswrapper[4742]: I1004 03:14:28.866249 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a" path="/var/lib/kubelet/pods/7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a/volumes" Oct 04 03:14:28 crc kubenswrapper[4742]: I1004 03:14:28.903191 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rd5tv"] Oct 04 03:14:28 crc kubenswrapper[4742]: I1004 03:14:28.903557 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rd5tv" podUID="3848a0f4-0c53-4a32-ad20-d4bce94d9a95" containerName="registry-server" containerID="cri-o://d7160de9da2516337c804dbbf975242d9409b13737e5d8c90eacbf552c9b5c85" gracePeriod=2 Oct 04 03:14:29 crc kubenswrapper[4742]: I1004 03:14:29.239355 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rd5tv" Oct 04 03:14:29 crc kubenswrapper[4742]: I1004 03:14:29.303256 4742 generic.go:334] "Generic (PLEG): container finished" podID="3848a0f4-0c53-4a32-ad20-d4bce94d9a95" containerID="d7160de9da2516337c804dbbf975242d9409b13737e5d8c90eacbf552c9b5c85" exitCode=0 Oct 04 03:14:29 crc kubenswrapper[4742]: I1004 03:14:29.303329 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rd5tv" Oct 04 03:14:29 crc kubenswrapper[4742]: I1004 03:14:29.303327 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rd5tv" event={"ID":"3848a0f4-0c53-4a32-ad20-d4bce94d9a95","Type":"ContainerDied","Data":"d7160de9da2516337c804dbbf975242d9409b13737e5d8c90eacbf552c9b5c85"} Oct 04 03:14:29 crc kubenswrapper[4742]: I1004 03:14:29.303465 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rd5tv" event={"ID":"3848a0f4-0c53-4a32-ad20-d4bce94d9a95","Type":"ContainerDied","Data":"f6535f2b3572d9958334a113f49ef0c9e2abb7419f6d5032e5041a4f5b7b9ebc"} Oct 04 03:14:29 crc kubenswrapper[4742]: I1004 03:14:29.303486 4742 scope.go:117] "RemoveContainer" containerID="d7160de9da2516337c804dbbf975242d9409b13737e5d8c90eacbf552c9b5c85" Oct 04 03:14:29 crc kubenswrapper[4742]: I1004 03:14:29.317710 4742 scope.go:117] "RemoveContainer" containerID="52f5eba945f2babfdf82d376ea6306dadcaeeb5263443e1e6afcfa6557f995c1" Oct 04 03:14:29 crc kubenswrapper[4742]: I1004 03:14:29.331561 4742 scope.go:117] "RemoveContainer" containerID="1eab366f2d8201e823eab2977bab4efc5034f1efa0376b8a0a06e2c7ccbcb184" Oct 04 03:14:29 crc kubenswrapper[4742]: I1004 03:14:29.346458 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3848a0f4-0c53-4a32-ad20-d4bce94d9a95-catalog-content\") pod \"3848a0f4-0c53-4a32-ad20-d4bce94d9a95\" (UID: \"3848a0f4-0c53-4a32-ad20-d4bce94d9a95\") " Oct 04 03:14:29 crc kubenswrapper[4742]: I1004 03:14:29.346501 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3848a0f4-0c53-4a32-ad20-d4bce94d9a95-utilities\") pod \"3848a0f4-0c53-4a32-ad20-d4bce94d9a95\" (UID: \"3848a0f4-0c53-4a32-ad20-d4bce94d9a95\") " Oct 04 03:14:29 crc kubenswrapper[4742]: I1004 03:14:29.346538 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzc98\" (UniqueName: \"kubernetes.io/projected/3848a0f4-0c53-4a32-ad20-d4bce94d9a95-kube-api-access-gzc98\") pod \"3848a0f4-0c53-4a32-ad20-d4bce94d9a95\" (UID: \"3848a0f4-0c53-4a32-ad20-d4bce94d9a95\") " Oct 04 03:14:29 crc kubenswrapper[4742]: I1004 03:14:29.347372 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3848a0f4-0c53-4a32-ad20-d4bce94d9a95-utilities" (OuterVolumeSpecName: "utilities") pod "3848a0f4-0c53-4a32-ad20-d4bce94d9a95" (UID: "3848a0f4-0c53-4a32-ad20-d4bce94d9a95"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:14:29 crc kubenswrapper[4742]: I1004 03:14:29.350529 4742 scope.go:117] "RemoveContainer" containerID="d7160de9da2516337c804dbbf975242d9409b13737e5d8c90eacbf552c9b5c85" Oct 04 03:14:29 crc kubenswrapper[4742]: E1004 03:14:29.351269 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7160de9da2516337c804dbbf975242d9409b13737e5d8c90eacbf552c9b5c85\": container with ID starting with d7160de9da2516337c804dbbf975242d9409b13737e5d8c90eacbf552c9b5c85 not found: ID does not exist" containerID="d7160de9da2516337c804dbbf975242d9409b13737e5d8c90eacbf552c9b5c85" Oct 04 03:14:29 crc kubenswrapper[4742]: I1004 03:14:29.351314 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7160de9da2516337c804dbbf975242d9409b13737e5d8c90eacbf552c9b5c85"} err="failed to get container status \"d7160de9da2516337c804dbbf975242d9409b13737e5d8c90eacbf552c9b5c85\": rpc error: code = NotFound desc = could not find container \"d7160de9da2516337c804dbbf975242d9409b13737e5d8c90eacbf552c9b5c85\": container with ID starting with d7160de9da2516337c804dbbf975242d9409b13737e5d8c90eacbf552c9b5c85 not found: ID does not exist" Oct 04 03:14:29 crc kubenswrapper[4742]: I1004 03:14:29.351337 4742 scope.go:117] "RemoveContainer" containerID="52f5eba945f2babfdf82d376ea6306dadcaeeb5263443e1e6afcfa6557f995c1" Oct 04 03:14:29 crc kubenswrapper[4742]: E1004 03:14:29.351827 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52f5eba945f2babfdf82d376ea6306dadcaeeb5263443e1e6afcfa6557f995c1\": container with ID starting with 52f5eba945f2babfdf82d376ea6306dadcaeeb5263443e1e6afcfa6557f995c1 not found: ID does not exist" containerID="52f5eba945f2babfdf82d376ea6306dadcaeeb5263443e1e6afcfa6557f995c1" Oct 04 03:14:29 crc kubenswrapper[4742]: I1004 03:14:29.351847 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52f5eba945f2babfdf82d376ea6306dadcaeeb5263443e1e6afcfa6557f995c1"} err="failed to get container status \"52f5eba945f2babfdf82d376ea6306dadcaeeb5263443e1e6afcfa6557f995c1\": rpc error: code = NotFound desc = could not find container \"52f5eba945f2babfdf82d376ea6306dadcaeeb5263443e1e6afcfa6557f995c1\": container with ID starting with 52f5eba945f2babfdf82d376ea6306dadcaeeb5263443e1e6afcfa6557f995c1 not found: ID does not exist" Oct 04 03:14:29 crc kubenswrapper[4742]: I1004 03:14:29.351861 4742 scope.go:117] "RemoveContainer" containerID="1eab366f2d8201e823eab2977bab4efc5034f1efa0376b8a0a06e2c7ccbcb184" Oct 04 03:14:29 crc kubenswrapper[4742]: E1004 03:14:29.352170 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1eab366f2d8201e823eab2977bab4efc5034f1efa0376b8a0a06e2c7ccbcb184\": container with ID starting with 1eab366f2d8201e823eab2977bab4efc5034f1efa0376b8a0a06e2c7ccbcb184 not found: ID does not exist" containerID="1eab366f2d8201e823eab2977bab4efc5034f1efa0376b8a0a06e2c7ccbcb184" Oct 04 03:14:29 crc kubenswrapper[4742]: I1004 03:14:29.352194 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1eab366f2d8201e823eab2977bab4efc5034f1efa0376b8a0a06e2c7ccbcb184"} err="failed to get container status \"1eab366f2d8201e823eab2977bab4efc5034f1efa0376b8a0a06e2c7ccbcb184\": rpc error: code = NotFound desc = could not find container \"1eab366f2d8201e823eab2977bab4efc5034f1efa0376b8a0a06e2c7ccbcb184\": container with ID starting with 1eab366f2d8201e823eab2977bab4efc5034f1efa0376b8a0a06e2c7ccbcb184 not found: ID does not exist" Oct 04 03:14:29 crc kubenswrapper[4742]: I1004 03:14:29.353227 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3848a0f4-0c53-4a32-ad20-d4bce94d9a95-kube-api-access-gzc98" (OuterVolumeSpecName: "kube-api-access-gzc98") pod "3848a0f4-0c53-4a32-ad20-d4bce94d9a95" (UID: "3848a0f4-0c53-4a32-ad20-d4bce94d9a95"). InnerVolumeSpecName "kube-api-access-gzc98". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:14:29 crc kubenswrapper[4742]: I1004 03:14:29.426854 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3848a0f4-0c53-4a32-ad20-d4bce94d9a95-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3848a0f4-0c53-4a32-ad20-d4bce94d9a95" (UID: "3848a0f4-0c53-4a32-ad20-d4bce94d9a95"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:14:29 crc kubenswrapper[4742]: I1004 03:14:29.447694 4742 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3848a0f4-0c53-4a32-ad20-d4bce94d9a95-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:29 crc kubenswrapper[4742]: I1004 03:14:29.447948 4742 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3848a0f4-0c53-4a32-ad20-d4bce94d9a95-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:29 crc kubenswrapper[4742]: I1004 03:14:29.448090 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzc98\" (UniqueName: \"kubernetes.io/projected/3848a0f4-0c53-4a32-ad20-d4bce94d9a95-kube-api-access-gzc98\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:29 crc kubenswrapper[4742]: I1004 03:14:29.626929 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rd5tv"] Oct 04 03:14:29 crc kubenswrapper[4742]: I1004 03:14:29.630951 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rd5tv"] Oct 04 03:14:30 crc kubenswrapper[4742]: I1004 03:14:30.865316 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3848a0f4-0c53-4a32-ad20-d4bce94d9a95" path="/var/lib/kubelet/pods/3848a0f4-0c53-4a32-ad20-d4bce94d9a95/volumes" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.073075 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" podUID="56722804-fa7c-4639-a4d3-5085cd965f8b" containerName="oauth-openshift" containerID="cri-o://cc9433abc93ed4eadb18b6ea683c8985ffe75711073814e6e91e1da79600208a" gracePeriod=15 Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.400826 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.416340 4742 generic.go:334] "Generic (PLEG): container finished" podID="56722804-fa7c-4639-a4d3-5085cd965f8b" containerID="cc9433abc93ed4eadb18b6ea683c8985ffe75711073814e6e91e1da79600208a" exitCode=0 Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.416393 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" event={"ID":"56722804-fa7c-4639-a4d3-5085cd965f8b","Type":"ContainerDied","Data":"cc9433abc93ed4eadb18b6ea683c8985ffe75711073814e6e91e1da79600208a"} Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.416414 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.416445 4742 scope.go:117] "RemoveContainer" containerID="cc9433abc93ed4eadb18b6ea683c8985ffe75711073814e6e91e1da79600208a" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.416430 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mspvb" event={"ID":"56722804-fa7c-4639-a4d3-5085cd965f8b","Type":"ContainerDied","Data":"b9ca38ac05caade3628be76b9f62478ef0999ffad533cf94b7cc6ff140a9c7bb"} Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.431704 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-57569d6b9d-w7s5w"] Oct 04 03:14:50 crc kubenswrapper[4742]: E1004 03:14:50.431905 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9afa0f84-44ab-47b1-986f-b545fddb3824" containerName="registry-server" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.431916 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="9afa0f84-44ab-47b1-986f-b545fddb3824" containerName="registry-server" Oct 04 03:14:50 crc kubenswrapper[4742]: E1004 03:14:50.431929 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a" containerName="registry-server" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.431935 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a" containerName="registry-server" Oct 04 03:14:50 crc kubenswrapper[4742]: E1004 03:14:50.431942 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afe664a2-9cfc-4a72-9903-08b2dbba5d3c" containerName="pruner" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.431948 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="afe664a2-9cfc-4a72-9903-08b2dbba5d3c" containerName="pruner" Oct 04 03:14:50 crc kubenswrapper[4742]: E1004 03:14:50.431954 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a" containerName="extract-content" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.431960 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a" containerName="extract-content" Oct 04 03:14:50 crc kubenswrapper[4742]: E1004 03:14:50.431969 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fbf09fa-a448-4ac9-8550-e9c719ec22d2" containerName="extract-utilities" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.431974 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fbf09fa-a448-4ac9-8550-e9c719ec22d2" containerName="extract-utilities" Oct 04 03:14:50 crc kubenswrapper[4742]: E1004 03:14:50.431982 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fbf09fa-a448-4ac9-8550-e9c719ec22d2" containerName="extract-content" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.431988 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fbf09fa-a448-4ac9-8550-e9c719ec22d2" containerName="extract-content" Oct 04 03:14:50 crc kubenswrapper[4742]: E1004 03:14:50.431999 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3848a0f4-0c53-4a32-ad20-d4bce94d9a95" containerName="registry-server" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.432004 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="3848a0f4-0c53-4a32-ad20-d4bce94d9a95" containerName="registry-server" Oct 04 03:14:50 crc kubenswrapper[4742]: E1004 03:14:50.432011 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a" containerName="extract-utilities" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.432017 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a" containerName="extract-utilities" Oct 04 03:14:50 crc kubenswrapper[4742]: E1004 03:14:50.432027 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56722804-fa7c-4639-a4d3-5085cd965f8b" containerName="oauth-openshift" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.432034 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="56722804-fa7c-4639-a4d3-5085cd965f8b" containerName="oauth-openshift" Oct 04 03:14:50 crc kubenswrapper[4742]: E1004 03:14:50.432043 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fbf09fa-a448-4ac9-8550-e9c719ec22d2" containerName="registry-server" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.432049 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fbf09fa-a448-4ac9-8550-e9c719ec22d2" containerName="registry-server" Oct 04 03:14:50 crc kubenswrapper[4742]: E1004 03:14:50.432059 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9afa0f84-44ab-47b1-986f-b545fddb3824" containerName="extract-utilities" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.432066 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="9afa0f84-44ab-47b1-986f-b545fddb3824" containerName="extract-utilities" Oct 04 03:14:50 crc kubenswrapper[4742]: E1004 03:14:50.432074 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3848a0f4-0c53-4a32-ad20-d4bce94d9a95" containerName="extract-utilities" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.432080 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="3848a0f4-0c53-4a32-ad20-d4bce94d9a95" containerName="extract-utilities" Oct 04 03:14:50 crc kubenswrapper[4742]: E1004 03:14:50.432090 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3848a0f4-0c53-4a32-ad20-d4bce94d9a95" containerName="extract-content" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.432097 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="3848a0f4-0c53-4a32-ad20-d4bce94d9a95" containerName="extract-content" Oct 04 03:14:50 crc kubenswrapper[4742]: E1004 03:14:50.432108 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9afa0f84-44ab-47b1-986f-b545fddb3824" containerName="extract-content" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.432115 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="9afa0f84-44ab-47b1-986f-b545fddb3824" containerName="extract-content" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.432222 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="9afa0f84-44ab-47b1-986f-b545fddb3824" containerName="registry-server" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.432234 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d987b94-a7f1-4172-a1e5-1cb56bfe7d4a" containerName="registry-server" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.432245 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="afe664a2-9cfc-4a72-9903-08b2dbba5d3c" containerName="pruner" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.432254 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fbf09fa-a448-4ac9-8550-e9c719ec22d2" containerName="registry-server" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.432297 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="3848a0f4-0c53-4a32-ad20-d4bce94d9a95" containerName="registry-server" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.432308 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="56722804-fa7c-4639-a4d3-5085cd965f8b" containerName="oauth-openshift" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.432697 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.444193 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-57569d6b9d-w7s5w"] Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.454547 4742 scope.go:117] "RemoveContainer" containerID="cc9433abc93ed4eadb18b6ea683c8985ffe75711073814e6e91e1da79600208a" Oct 04 03:14:50 crc kubenswrapper[4742]: E1004 03:14:50.456447 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc9433abc93ed4eadb18b6ea683c8985ffe75711073814e6e91e1da79600208a\": container with ID starting with cc9433abc93ed4eadb18b6ea683c8985ffe75711073814e6e91e1da79600208a not found: ID does not exist" containerID="cc9433abc93ed4eadb18b6ea683c8985ffe75711073814e6e91e1da79600208a" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.456503 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc9433abc93ed4eadb18b6ea683c8985ffe75711073814e6e91e1da79600208a"} err="failed to get container status \"cc9433abc93ed4eadb18b6ea683c8985ffe75711073814e6e91e1da79600208a\": rpc error: code = NotFound desc = could not find container \"cc9433abc93ed4eadb18b6ea683c8985ffe75711073814e6e91e1da79600208a\": container with ID starting with cc9433abc93ed4eadb18b6ea683c8985ffe75711073814e6e91e1da79600208a not found: ID does not exist" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.489349 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-user-template-login\") pod \"56722804-fa7c-4639-a4d3-5085cd965f8b\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.489411 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-session\") pod \"56722804-fa7c-4639-a4d3-5085cd965f8b\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.489434 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/56722804-fa7c-4639-a4d3-5085cd965f8b-audit-policies\") pod \"56722804-fa7c-4639-a4d3-5085cd965f8b\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.489459 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-ocp-branding-template\") pod \"56722804-fa7c-4639-a4d3-5085cd965f8b\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.489490 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-user-template-provider-selection\") pod \"56722804-fa7c-4639-a4d3-5085cd965f8b\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.489509 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-service-ca\") pod \"56722804-fa7c-4639-a4d3-5085cd965f8b\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.489527 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9w5jf\" (UniqueName: \"kubernetes.io/projected/56722804-fa7c-4639-a4d3-5085cd965f8b-kube-api-access-9w5jf\") pod \"56722804-fa7c-4639-a4d3-5085cd965f8b\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.489549 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-user-template-error\") pod \"56722804-fa7c-4639-a4d3-5085cd965f8b\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.489608 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-cliconfig\") pod \"56722804-fa7c-4639-a4d3-5085cd965f8b\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.489626 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-serving-cert\") pod \"56722804-fa7c-4639-a4d3-5085cd965f8b\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.489644 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-user-idp-0-file-data\") pod \"56722804-fa7c-4639-a4d3-5085cd965f8b\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.489665 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/56722804-fa7c-4639-a4d3-5085cd965f8b-audit-dir\") pod \"56722804-fa7c-4639-a4d3-5085cd965f8b\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.489683 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-trusted-ca-bundle\") pod \"56722804-fa7c-4639-a4d3-5085cd965f8b\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.489702 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-router-certs\") pod \"56722804-fa7c-4639-a4d3-5085cd965f8b\" (UID: \"56722804-fa7c-4639-a4d3-5085cd965f8b\") " Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.490567 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/56722804-fa7c-4639-a4d3-5085cd965f8b-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "56722804-fa7c-4639-a4d3-5085cd965f8b" (UID: "56722804-fa7c-4639-a4d3-5085cd965f8b"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.491193 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c4e103a4-149e-4118-848a-5d22f9ef507f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.491255 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c4e103a4-149e-4118-848a-5d22f9ef507f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.491327 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c4e103a4-149e-4118-848a-5d22f9ef507f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.491361 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c4e103a4-149e-4118-848a-5d22f9ef507f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.491413 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c4e103a4-149e-4118-848a-5d22f9ef507f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.491450 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c4e103a4-149e-4118-848a-5d22f9ef507f-v4-0-config-system-router-certs\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.491458 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "56722804-fa7c-4639-a4d3-5085cd965f8b" (UID: "56722804-fa7c-4639-a4d3-5085cd965f8b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.491485 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c4e103a4-149e-4118-848a-5d22f9ef507f-v4-0-config-user-template-login\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.491514 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c4e103a4-149e-4118-848a-5d22f9ef507f-audit-dir\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.491560 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c4e103a4-149e-4118-848a-5d22f9ef507f-v4-0-config-system-session\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.491575 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56722804-fa7c-4639-a4d3-5085cd965f8b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "56722804-fa7c-4639-a4d3-5085cd965f8b" (UID: "56722804-fa7c-4639-a4d3-5085cd965f8b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.491590 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c4e103a4-149e-4118-848a-5d22f9ef507f-v4-0-config-system-service-ca\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.491615 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c4e103a4-149e-4118-848a-5d22f9ef507f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.491775 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbr2h\" (UniqueName: \"kubernetes.io/projected/c4e103a4-149e-4118-848a-5d22f9ef507f-kube-api-access-nbr2h\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.491846 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c4e103a4-149e-4118-848a-5d22f9ef507f-audit-policies\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.491918 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c4e103a4-149e-4118-848a-5d22f9ef507f-v4-0-config-user-template-error\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.491940 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "56722804-fa7c-4639-a4d3-5085cd965f8b" (UID: "56722804-fa7c-4639-a4d3-5085cd965f8b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.491965 4742 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/56722804-fa7c-4639-a4d3-5085cd965f8b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.492006 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "56722804-fa7c-4639-a4d3-5085cd965f8b" (UID: "56722804-fa7c-4639-a4d3-5085cd965f8b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.492022 4742 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/56722804-fa7c-4639-a4d3-5085cd965f8b-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.492045 4742 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.496384 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56722804-fa7c-4639-a4d3-5085cd965f8b-kube-api-access-9w5jf" (OuterVolumeSpecName: "kube-api-access-9w5jf") pod "56722804-fa7c-4639-a4d3-5085cd965f8b" (UID: "56722804-fa7c-4639-a4d3-5085cd965f8b"). InnerVolumeSpecName "kube-api-access-9w5jf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.496464 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "56722804-fa7c-4639-a4d3-5085cd965f8b" (UID: "56722804-fa7c-4639-a4d3-5085cd965f8b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.497563 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "56722804-fa7c-4639-a4d3-5085cd965f8b" (UID: "56722804-fa7c-4639-a4d3-5085cd965f8b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.498322 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "56722804-fa7c-4639-a4d3-5085cd965f8b" (UID: "56722804-fa7c-4639-a4d3-5085cd965f8b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.499212 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "56722804-fa7c-4639-a4d3-5085cd965f8b" (UID: "56722804-fa7c-4639-a4d3-5085cd965f8b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.500676 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "56722804-fa7c-4639-a4d3-5085cd965f8b" (UID: "56722804-fa7c-4639-a4d3-5085cd965f8b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.504768 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "56722804-fa7c-4639-a4d3-5085cd965f8b" (UID: "56722804-fa7c-4639-a4d3-5085cd965f8b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.510482 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "56722804-fa7c-4639-a4d3-5085cd965f8b" (UID: "56722804-fa7c-4639-a4d3-5085cd965f8b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.510516 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "56722804-fa7c-4639-a4d3-5085cd965f8b" (UID: "56722804-fa7c-4639-a4d3-5085cd965f8b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.592489 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c4e103a4-149e-4118-848a-5d22f9ef507f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.592545 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c4e103a4-149e-4118-848a-5d22f9ef507f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.592566 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c4e103a4-149e-4118-848a-5d22f9ef507f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.592585 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c4e103a4-149e-4118-848a-5d22f9ef507f-v4-0-config-system-router-certs\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.592604 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c4e103a4-149e-4118-848a-5d22f9ef507f-v4-0-config-user-template-login\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.592626 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c4e103a4-149e-4118-848a-5d22f9ef507f-audit-dir\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.592666 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c4e103a4-149e-4118-848a-5d22f9ef507f-v4-0-config-system-session\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.592686 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c4e103a4-149e-4118-848a-5d22f9ef507f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.592705 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c4e103a4-149e-4118-848a-5d22f9ef507f-v4-0-config-system-service-ca\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.592721 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbr2h\" (UniqueName: \"kubernetes.io/projected/c4e103a4-149e-4118-848a-5d22f9ef507f-kube-api-access-nbr2h\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.592742 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c4e103a4-149e-4118-848a-5d22f9ef507f-audit-policies\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.592778 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c4e103a4-149e-4118-848a-5d22f9ef507f-v4-0-config-user-template-error\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.592805 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c4e103a4-149e-4118-848a-5d22f9ef507f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.592824 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c4e103a4-149e-4118-848a-5d22f9ef507f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.592862 4742 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.592874 4742 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.592884 4742 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.592895 4742 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.592906 4742 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.592916 4742 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.592926 4742 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.592936 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9w5jf\" (UniqueName: \"kubernetes.io/projected/56722804-fa7c-4639-a4d3-5085cd965f8b-kube-api-access-9w5jf\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.592945 4742 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.592954 4742 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.592963 4742 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/56722804-fa7c-4639-a4d3-5085cd965f8b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.593835 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c4e103a4-149e-4118-848a-5d22f9ef507f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.594252 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c4e103a4-149e-4118-848a-5d22f9ef507f-audit-dir\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.594262 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c4e103a4-149e-4118-848a-5d22f9ef507f-audit-policies\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.594716 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c4e103a4-149e-4118-848a-5d22f9ef507f-v4-0-config-system-service-ca\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.594780 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c4e103a4-149e-4118-848a-5d22f9ef507f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.596735 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c4e103a4-149e-4118-848a-5d22f9ef507f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.596935 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c4e103a4-149e-4118-848a-5d22f9ef507f-v4-0-config-user-template-login\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.597157 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c4e103a4-149e-4118-848a-5d22f9ef507f-v4-0-config-system-session\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.598218 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c4e103a4-149e-4118-848a-5d22f9ef507f-v4-0-config-system-router-certs\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.598398 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c4e103a4-149e-4118-848a-5d22f9ef507f-v4-0-config-user-template-error\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.598731 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c4e103a4-149e-4118-848a-5d22f9ef507f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.599062 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c4e103a4-149e-4118-848a-5d22f9ef507f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.599508 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c4e103a4-149e-4118-848a-5d22f9ef507f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.609054 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbr2h\" (UniqueName: \"kubernetes.io/projected/c4e103a4-149e-4118-848a-5d22f9ef507f-kube-api-access-nbr2h\") pod \"oauth-openshift-57569d6b9d-w7s5w\" (UID: \"c4e103a4-149e-4118-848a-5d22f9ef507f\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.743731 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mspvb"] Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.749072 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mspvb"] Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.756062 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.865385 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56722804-fa7c-4639-a4d3-5085cd965f8b" path="/var/lib/kubelet/pods/56722804-fa7c-4639-a4d3-5085cd965f8b/volumes" Oct 04 03:14:50 crc kubenswrapper[4742]: I1004 03:14:50.945291 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-57569d6b9d-w7s5w"] Oct 04 03:14:51 crc kubenswrapper[4742]: I1004 03:14:51.426334 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" event={"ID":"c4e103a4-149e-4118-848a-5d22f9ef507f","Type":"ContainerStarted","Data":"a3358852b4456b26f6a0b419766b28f735c4dba6d403df607cbb854bfa335bef"} Oct 04 03:14:51 crc kubenswrapper[4742]: I1004 03:14:51.426386 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" event={"ID":"c4e103a4-149e-4118-848a-5d22f9ef507f","Type":"ContainerStarted","Data":"e98adddab0ba9e9de080a5a53de50fa0429f8c47275e645a98493b2387a7321d"} Oct 04 03:14:51 crc kubenswrapper[4742]: I1004 03:14:51.427190 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:14:51 crc kubenswrapper[4742]: I1004 03:14:51.451867 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" podStartSLOduration=26.451840888 podStartE2EDuration="26.451840888s" podCreationTimestamp="2025-10-04 03:14:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:14:51.445207433 +0000 UTC m=+257.377544609" watchObservedRunningTime="2025-10-04 03:14:51.451840888 +0000 UTC m=+257.384178074" Oct 04 03:14:51 crc kubenswrapper[4742]: I1004 03:14:51.668422 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-57569d6b9d-w7s5w" Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.131062 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325795-vhpcq"] Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.132300 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325795-vhpcq" Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.135982 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.136171 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.143729 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325795-vhpcq"] Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.315602 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/71e52223-cb12-4aae-8dcf-7418f39073c7-config-volume\") pod \"collect-profiles-29325795-vhpcq\" (UID: \"71e52223-cb12-4aae-8dcf-7418f39073c7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325795-vhpcq" Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.315872 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hp5q6\" (UniqueName: \"kubernetes.io/projected/71e52223-cb12-4aae-8dcf-7418f39073c7-kube-api-access-hp5q6\") pod \"collect-profiles-29325795-vhpcq\" (UID: \"71e52223-cb12-4aae-8dcf-7418f39073c7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325795-vhpcq" Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.315951 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/71e52223-cb12-4aae-8dcf-7418f39073c7-secret-volume\") pod \"collect-profiles-29325795-vhpcq\" (UID: \"71e52223-cb12-4aae-8dcf-7418f39073c7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325795-vhpcq" Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.417050 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/71e52223-cb12-4aae-8dcf-7418f39073c7-config-volume\") pod \"collect-profiles-29325795-vhpcq\" (UID: \"71e52223-cb12-4aae-8dcf-7418f39073c7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325795-vhpcq" Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.417137 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hp5q6\" (UniqueName: \"kubernetes.io/projected/71e52223-cb12-4aae-8dcf-7418f39073c7-kube-api-access-hp5q6\") pod \"collect-profiles-29325795-vhpcq\" (UID: \"71e52223-cb12-4aae-8dcf-7418f39073c7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325795-vhpcq" Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.417223 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/71e52223-cb12-4aae-8dcf-7418f39073c7-secret-volume\") pod \"collect-profiles-29325795-vhpcq\" (UID: \"71e52223-cb12-4aae-8dcf-7418f39073c7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325795-vhpcq" Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.417914 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/71e52223-cb12-4aae-8dcf-7418f39073c7-config-volume\") pod \"collect-profiles-29325795-vhpcq\" (UID: \"71e52223-cb12-4aae-8dcf-7418f39073c7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325795-vhpcq" Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.442800 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/71e52223-cb12-4aae-8dcf-7418f39073c7-secret-volume\") pod \"collect-profiles-29325795-vhpcq\" (UID: \"71e52223-cb12-4aae-8dcf-7418f39073c7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325795-vhpcq" Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.446997 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hp5q6\" (UniqueName: \"kubernetes.io/projected/71e52223-cb12-4aae-8dcf-7418f39073c7-kube-api-access-hp5q6\") pod \"collect-profiles-29325795-vhpcq\" (UID: \"71e52223-cb12-4aae-8dcf-7418f39073c7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325795-vhpcq" Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.461887 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325795-vhpcq" Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.614186 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b5xzp"] Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.614721 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-b5xzp" podUID="f3b8a6f1-47db-4541-af94-5519b15d1407" containerName="registry-server" containerID="cri-o://052b37f81317c68a03ea64f3337d7574c49e74724e2c21557dabc01db252e9aa" gracePeriod=30 Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.634370 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-trfnh"] Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.634701 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-trfnh" podUID="2788ab92-5c72-4f27-888e-ef695e4cf6cb" containerName="registry-server" containerID="cri-o://abaa6ec10825e7beaa1a4d045a9d20823ba519919ca5549d131d09c1fb50dee6" gracePeriod=30 Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.647254 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tfc7g"] Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.650172 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-tfc7g" podUID="a8b41c08-4e89-4f08-8d45-256c306fbea0" containerName="marketplace-operator" containerID="cri-o://9f9c46a4ab5b18d70d78e1592b5d352927e2103c6d485583eb027a29fe3d3fa5" gracePeriod=30 Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.650389 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325795-vhpcq"] Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.655918 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lb54b"] Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.656241 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lb54b" podUID="c3fa495f-e19d-4658-a335-1e49cb14a721" containerName="registry-server" containerID="cri-o://4acb3da1f52b48c905c1821afc3f6102d530d39dfc73cc8b87ba3ef724b85dff" gracePeriod=30 Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.666703 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4mmhw"] Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.667680 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4mmhw" Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.677538 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-67ztt"] Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.677783 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-67ztt" podUID="c210acab-8945-4f9d-a8fe-47ae7d78a064" containerName="registry-server" containerID="cri-o://dafdecd26505ab05deade27027f7e144160fa80a5642e21b0fcc722372082dcd" gracePeriod=30 Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.685087 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4mmhw"] Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.820386 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/18ed4cd4-2848-4f6f-bf76-bd1425aa40b5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4mmhw\" (UID: \"18ed4cd4-2848-4f6f-bf76-bd1425aa40b5\") " pod="openshift-marketplace/marketplace-operator-79b997595-4mmhw" Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.820464 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/18ed4cd4-2848-4f6f-bf76-bd1425aa40b5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4mmhw\" (UID: \"18ed4cd4-2848-4f6f-bf76-bd1425aa40b5\") " pod="openshift-marketplace/marketplace-operator-79b997595-4mmhw" Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.820494 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcq4z\" (UniqueName: \"kubernetes.io/projected/18ed4cd4-2848-4f6f-bf76-bd1425aa40b5-kube-api-access-xcq4z\") pod \"marketplace-operator-79b997595-4mmhw\" (UID: \"18ed4cd4-2848-4f6f-bf76-bd1425aa40b5\") " pod="openshift-marketplace/marketplace-operator-79b997595-4mmhw" Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.921919 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/18ed4cd4-2848-4f6f-bf76-bd1425aa40b5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4mmhw\" (UID: \"18ed4cd4-2848-4f6f-bf76-bd1425aa40b5\") " pod="openshift-marketplace/marketplace-operator-79b997595-4mmhw" Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.921990 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/18ed4cd4-2848-4f6f-bf76-bd1425aa40b5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4mmhw\" (UID: \"18ed4cd4-2848-4f6f-bf76-bd1425aa40b5\") " pod="openshift-marketplace/marketplace-operator-79b997595-4mmhw" Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.922021 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcq4z\" (UniqueName: \"kubernetes.io/projected/18ed4cd4-2848-4f6f-bf76-bd1425aa40b5-kube-api-access-xcq4z\") pod \"marketplace-operator-79b997595-4mmhw\" (UID: \"18ed4cd4-2848-4f6f-bf76-bd1425aa40b5\") " pod="openshift-marketplace/marketplace-operator-79b997595-4mmhw" Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.924715 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/18ed4cd4-2848-4f6f-bf76-bd1425aa40b5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4mmhw\" (UID: \"18ed4cd4-2848-4f6f-bf76-bd1425aa40b5\") " pod="openshift-marketplace/marketplace-operator-79b997595-4mmhw" Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.928909 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/18ed4cd4-2848-4f6f-bf76-bd1425aa40b5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4mmhw\" (UID: \"18ed4cd4-2848-4f6f-bf76-bd1425aa40b5\") " pod="openshift-marketplace/marketplace-operator-79b997595-4mmhw" Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.939479 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcq4z\" (UniqueName: \"kubernetes.io/projected/18ed4cd4-2848-4f6f-bf76-bd1425aa40b5-kube-api-access-xcq4z\") pod \"marketplace-operator-79b997595-4mmhw\" (UID: \"18ed4cd4-2848-4f6f-bf76-bd1425aa40b5\") " pod="openshift-marketplace/marketplace-operator-79b997595-4mmhw" Oct 04 03:15:00 crc kubenswrapper[4742]: I1004 03:15:00.995135 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4mmhw" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.010869 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-trfnh" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.046841 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-tfc7g" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.052179 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lb54b" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.132429 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2788ab92-5c72-4f27-888e-ef695e4cf6cb-catalog-content\") pod \"2788ab92-5c72-4f27-888e-ef695e4cf6cb\" (UID: \"2788ab92-5c72-4f27-888e-ef695e4cf6cb\") " Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.132505 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zc4gb\" (UniqueName: \"kubernetes.io/projected/2788ab92-5c72-4f27-888e-ef695e4cf6cb-kube-api-access-zc4gb\") pod \"2788ab92-5c72-4f27-888e-ef695e4cf6cb\" (UID: \"2788ab92-5c72-4f27-888e-ef695e4cf6cb\") " Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.132539 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2788ab92-5c72-4f27-888e-ef695e4cf6cb-utilities\") pod \"2788ab92-5c72-4f27-888e-ef695e4cf6cb\" (UID: \"2788ab92-5c72-4f27-888e-ef695e4cf6cb\") " Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.135595 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2788ab92-5c72-4f27-888e-ef695e4cf6cb-utilities" (OuterVolumeSpecName: "utilities") pod "2788ab92-5c72-4f27-888e-ef695e4cf6cb" (UID: "2788ab92-5c72-4f27-888e-ef695e4cf6cb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.136255 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2788ab92-5c72-4f27-888e-ef695e4cf6cb-kube-api-access-zc4gb" (OuterVolumeSpecName: "kube-api-access-zc4gb") pod "2788ab92-5c72-4f27-888e-ef695e4cf6cb" (UID: "2788ab92-5c72-4f27-888e-ef695e4cf6cb"). InnerVolumeSpecName "kube-api-access-zc4gb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.148625 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-67ztt" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.209420 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2788ab92-5c72-4f27-888e-ef695e4cf6cb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2788ab92-5c72-4f27-888e-ef695e4cf6cb" (UID: "2788ab92-5c72-4f27-888e-ef695e4cf6cb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.234800 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x55xr\" (UniqueName: \"kubernetes.io/projected/a8b41c08-4e89-4f08-8d45-256c306fbea0-kube-api-access-x55xr\") pod \"a8b41c08-4e89-4f08-8d45-256c306fbea0\" (UID: \"a8b41c08-4e89-4f08-8d45-256c306fbea0\") " Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.234860 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3fa495f-e19d-4658-a335-1e49cb14a721-catalog-content\") pod \"c3fa495f-e19d-4658-a335-1e49cb14a721\" (UID: \"c3fa495f-e19d-4658-a335-1e49cb14a721\") " Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.234930 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3fa495f-e19d-4658-a335-1e49cb14a721-utilities\") pod \"c3fa495f-e19d-4658-a335-1e49cb14a721\" (UID: \"c3fa495f-e19d-4658-a335-1e49cb14a721\") " Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.234973 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a8b41c08-4e89-4f08-8d45-256c306fbea0-marketplace-trusted-ca\") pod \"a8b41c08-4e89-4f08-8d45-256c306fbea0\" (UID: \"a8b41c08-4e89-4f08-8d45-256c306fbea0\") " Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.234992 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a8b41c08-4e89-4f08-8d45-256c306fbea0-marketplace-operator-metrics\") pod \"a8b41c08-4e89-4f08-8d45-256c306fbea0\" (UID: \"a8b41c08-4e89-4f08-8d45-256c306fbea0\") " Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.235159 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7jjg\" (UniqueName: \"kubernetes.io/projected/c3fa495f-e19d-4658-a335-1e49cb14a721-kube-api-access-k7jjg\") pod \"c3fa495f-e19d-4658-a335-1e49cb14a721\" (UID: \"c3fa495f-e19d-4658-a335-1e49cb14a721\") " Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.236423 4742 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2788ab92-5c72-4f27-888e-ef695e4cf6cb-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.236453 4742 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2788ab92-5c72-4f27-888e-ef695e4cf6cb-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.236466 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zc4gb\" (UniqueName: \"kubernetes.io/projected/2788ab92-5c72-4f27-888e-ef695e4cf6cb-kube-api-access-zc4gb\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.237090 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3fa495f-e19d-4658-a335-1e49cb14a721-utilities" (OuterVolumeSpecName: "utilities") pod "c3fa495f-e19d-4658-a335-1e49cb14a721" (UID: "c3fa495f-e19d-4658-a335-1e49cb14a721"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.237520 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8b41c08-4e89-4f08-8d45-256c306fbea0-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "a8b41c08-4e89-4f08-8d45-256c306fbea0" (UID: "a8b41c08-4e89-4f08-8d45-256c306fbea0"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.239224 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8b41c08-4e89-4f08-8d45-256c306fbea0-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "a8b41c08-4e89-4f08-8d45-256c306fbea0" (UID: "a8b41c08-4e89-4f08-8d45-256c306fbea0"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.239556 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3fa495f-e19d-4658-a335-1e49cb14a721-kube-api-access-k7jjg" (OuterVolumeSpecName: "kube-api-access-k7jjg") pod "c3fa495f-e19d-4658-a335-1e49cb14a721" (UID: "c3fa495f-e19d-4658-a335-1e49cb14a721"). InnerVolumeSpecName "kube-api-access-k7jjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.241212 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8b41c08-4e89-4f08-8d45-256c306fbea0-kube-api-access-x55xr" (OuterVolumeSpecName: "kube-api-access-x55xr") pod "a8b41c08-4e89-4f08-8d45-256c306fbea0" (UID: "a8b41c08-4e89-4f08-8d45-256c306fbea0"). InnerVolumeSpecName "kube-api-access-x55xr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.248962 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3fa495f-e19d-4658-a335-1e49cb14a721-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c3fa495f-e19d-4658-a335-1e49cb14a721" (UID: "c3fa495f-e19d-4658-a335-1e49cb14a721"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.251515 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b5xzp" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.288693 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4mmhw"] Oct 04 03:15:01 crc kubenswrapper[4742]: W1004 03:15:01.293346 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18ed4cd4_2848_4f6f_bf76_bd1425aa40b5.slice/crio-ecc0e8550cb8df3a6da6a427595d91af43c71658f56e68c390acb4d21d2779a5 WatchSource:0}: Error finding container ecc0e8550cb8df3a6da6a427595d91af43c71658f56e68c390acb4d21d2779a5: Status 404 returned error can't find the container with id ecc0e8550cb8df3a6da6a427595d91af43c71658f56e68c390acb4d21d2779a5 Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.337267 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3b8a6f1-47db-4541-af94-5519b15d1407-catalog-content\") pod \"f3b8a6f1-47db-4541-af94-5519b15d1407\" (UID: \"f3b8a6f1-47db-4541-af94-5519b15d1407\") " Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.337364 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nj7w\" (UniqueName: \"kubernetes.io/projected/c210acab-8945-4f9d-a8fe-47ae7d78a064-kube-api-access-2nj7w\") pod \"c210acab-8945-4f9d-a8fe-47ae7d78a064\" (UID: \"c210acab-8945-4f9d-a8fe-47ae7d78a064\") " Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.337392 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3b8a6f1-47db-4541-af94-5519b15d1407-utilities\") pod \"f3b8a6f1-47db-4541-af94-5519b15d1407\" (UID: \"f3b8a6f1-47db-4541-af94-5519b15d1407\") " Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.337410 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c210acab-8945-4f9d-a8fe-47ae7d78a064-utilities\") pod \"c210acab-8945-4f9d-a8fe-47ae7d78a064\" (UID: \"c210acab-8945-4f9d-a8fe-47ae7d78a064\") " Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.337430 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dcmvr\" (UniqueName: \"kubernetes.io/projected/f3b8a6f1-47db-4541-af94-5519b15d1407-kube-api-access-dcmvr\") pod \"f3b8a6f1-47db-4541-af94-5519b15d1407\" (UID: \"f3b8a6f1-47db-4541-af94-5519b15d1407\") " Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.337479 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c210acab-8945-4f9d-a8fe-47ae7d78a064-catalog-content\") pod \"c210acab-8945-4f9d-a8fe-47ae7d78a064\" (UID: \"c210acab-8945-4f9d-a8fe-47ae7d78a064\") " Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.337703 4742 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3fa495f-e19d-4658-a335-1e49cb14a721-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.337717 4742 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3fa495f-e19d-4658-a335-1e49cb14a721-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.337726 4742 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a8b41c08-4e89-4f08-8d45-256c306fbea0-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.337738 4742 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a8b41c08-4e89-4f08-8d45-256c306fbea0-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.337747 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7jjg\" (UniqueName: \"kubernetes.io/projected/c3fa495f-e19d-4658-a335-1e49cb14a721-kube-api-access-k7jjg\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.337755 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x55xr\" (UniqueName: \"kubernetes.io/projected/a8b41c08-4e89-4f08-8d45-256c306fbea0-kube-api-access-x55xr\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.338198 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3b8a6f1-47db-4541-af94-5519b15d1407-utilities" (OuterVolumeSpecName: "utilities") pod "f3b8a6f1-47db-4541-af94-5519b15d1407" (UID: "f3b8a6f1-47db-4541-af94-5519b15d1407"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.339083 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c210acab-8945-4f9d-a8fe-47ae7d78a064-utilities" (OuterVolumeSpecName: "utilities") pod "c210acab-8945-4f9d-a8fe-47ae7d78a064" (UID: "c210acab-8945-4f9d-a8fe-47ae7d78a064"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.341226 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c210acab-8945-4f9d-a8fe-47ae7d78a064-kube-api-access-2nj7w" (OuterVolumeSpecName: "kube-api-access-2nj7w") pod "c210acab-8945-4f9d-a8fe-47ae7d78a064" (UID: "c210acab-8945-4f9d-a8fe-47ae7d78a064"). InnerVolumeSpecName "kube-api-access-2nj7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.341564 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3b8a6f1-47db-4541-af94-5519b15d1407-kube-api-access-dcmvr" (OuterVolumeSpecName: "kube-api-access-dcmvr") pod "f3b8a6f1-47db-4541-af94-5519b15d1407" (UID: "f3b8a6f1-47db-4541-af94-5519b15d1407"). InnerVolumeSpecName "kube-api-access-dcmvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.388817 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3b8a6f1-47db-4541-af94-5519b15d1407-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f3b8a6f1-47db-4541-af94-5519b15d1407" (UID: "f3b8a6f1-47db-4541-af94-5519b15d1407"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.421791 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c210acab-8945-4f9d-a8fe-47ae7d78a064-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c210acab-8945-4f9d-a8fe-47ae7d78a064" (UID: "c210acab-8945-4f9d-a8fe-47ae7d78a064"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.438595 4742 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3b8a6f1-47db-4541-af94-5519b15d1407-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.438639 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nj7w\" (UniqueName: \"kubernetes.io/projected/c210acab-8945-4f9d-a8fe-47ae7d78a064-kube-api-access-2nj7w\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.438652 4742 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3b8a6f1-47db-4541-af94-5519b15d1407-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.438663 4742 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c210acab-8945-4f9d-a8fe-47ae7d78a064-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.438674 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dcmvr\" (UniqueName: \"kubernetes.io/projected/f3b8a6f1-47db-4541-af94-5519b15d1407-kube-api-access-dcmvr\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.438684 4742 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c210acab-8945-4f9d-a8fe-47ae7d78a064-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.483846 4742 generic.go:334] "Generic (PLEG): container finished" podID="c210acab-8945-4f9d-a8fe-47ae7d78a064" containerID="dafdecd26505ab05deade27027f7e144160fa80a5642e21b0fcc722372082dcd" exitCode=0 Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.483920 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-67ztt" event={"ID":"c210acab-8945-4f9d-a8fe-47ae7d78a064","Type":"ContainerDied","Data":"dafdecd26505ab05deade27027f7e144160fa80a5642e21b0fcc722372082dcd"} Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.483932 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-67ztt" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.483957 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-67ztt" event={"ID":"c210acab-8945-4f9d-a8fe-47ae7d78a064","Type":"ContainerDied","Data":"eb25dc8b6ffd2d353caddfd5fe77515a24b7fe51a4ba4beab8250b748b804d9c"} Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.483986 4742 scope.go:117] "RemoveContainer" containerID="dafdecd26505ab05deade27027f7e144160fa80a5642e21b0fcc722372082dcd" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.487821 4742 generic.go:334] "Generic (PLEG): container finished" podID="c3fa495f-e19d-4658-a335-1e49cb14a721" containerID="4acb3da1f52b48c905c1821afc3f6102d530d39dfc73cc8b87ba3ef724b85dff" exitCode=0 Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.487884 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lb54b" event={"ID":"c3fa495f-e19d-4658-a335-1e49cb14a721","Type":"ContainerDied","Data":"4acb3da1f52b48c905c1821afc3f6102d530d39dfc73cc8b87ba3ef724b85dff"} Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.487909 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lb54b" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.487913 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lb54b" event={"ID":"c3fa495f-e19d-4658-a335-1e49cb14a721","Type":"ContainerDied","Data":"db7fa3e13b3a8fa3e4ca52d2b1ff0eaf0ac9d4bb82ed7cd2fa3f6ea7e46eb2b4"} Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.490457 4742 generic.go:334] "Generic (PLEG): container finished" podID="2788ab92-5c72-4f27-888e-ef695e4cf6cb" containerID="abaa6ec10825e7beaa1a4d045a9d20823ba519919ca5549d131d09c1fb50dee6" exitCode=0 Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.490512 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-trfnh" event={"ID":"2788ab92-5c72-4f27-888e-ef695e4cf6cb","Type":"ContainerDied","Data":"abaa6ec10825e7beaa1a4d045a9d20823ba519919ca5549d131d09c1fb50dee6"} Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.490530 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-trfnh" event={"ID":"2788ab92-5c72-4f27-888e-ef695e4cf6cb","Type":"ContainerDied","Data":"33984713aca33cb6b6070d96e5726eebb5378c76d2d78f8d66f09542b6ea47c0"} Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.490586 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-trfnh" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.496385 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4mmhw" event={"ID":"18ed4cd4-2848-4f6f-bf76-bd1425aa40b5","Type":"ContainerStarted","Data":"84029bc22d338881eb55541fb2d2a9da073013e17d276c5b0b87f108a78595d1"} Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.496441 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4mmhw" event={"ID":"18ed4cd4-2848-4f6f-bf76-bd1425aa40b5","Type":"ContainerStarted","Data":"ecc0e8550cb8df3a6da6a427595d91af43c71658f56e68c390acb4d21d2779a5"} Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.496913 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-4mmhw" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.498346 4742 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-4mmhw container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.56:8080/healthz\": dial tcp 10.217.0.56:8080: connect: connection refused" start-of-body= Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.498453 4742 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-4mmhw" podUID="18ed4cd4-2848-4f6f-bf76-bd1425aa40b5" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.56:8080/healthz\": dial tcp 10.217.0.56:8080: connect: connection refused" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.513712 4742 generic.go:334] "Generic (PLEG): container finished" podID="71e52223-cb12-4aae-8dcf-7418f39073c7" containerID="1eb80461615f507ecf9de561b3982f09bd54f974d039ec8785fa9ef33ef3757b" exitCode=0 Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.513888 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325795-vhpcq" event={"ID":"71e52223-cb12-4aae-8dcf-7418f39073c7","Type":"ContainerDied","Data":"1eb80461615f507ecf9de561b3982f09bd54f974d039ec8785fa9ef33ef3757b"} Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.513927 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325795-vhpcq" event={"ID":"71e52223-cb12-4aae-8dcf-7418f39073c7","Type":"ContainerStarted","Data":"a0071d46424c0640c2bec58fa70821a3b772e95d24849d8bd8f93f335138fc9f"} Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.514445 4742 scope.go:117] "RemoveContainer" containerID="a0b830382262c3de530a309506e7061b48f3467ae89d38d2e2ffca8b801be0ea" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.522655 4742 generic.go:334] "Generic (PLEG): container finished" podID="a8b41c08-4e89-4f08-8d45-256c306fbea0" containerID="9f9c46a4ab5b18d70d78e1592b5d352927e2103c6d485583eb027a29fe3d3fa5" exitCode=0 Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.523507 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tfc7g" event={"ID":"a8b41c08-4e89-4f08-8d45-256c306fbea0","Type":"ContainerDied","Data":"9f9c46a4ab5b18d70d78e1592b5d352927e2103c6d485583eb027a29fe3d3fa5"} Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.523555 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tfc7g" event={"ID":"a8b41c08-4e89-4f08-8d45-256c306fbea0","Type":"ContainerDied","Data":"df9ce9e210f6d14e489c7061ca2d98d8972dc36b37ba107b79b69f93f09038ec"} Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.523985 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-tfc7g" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.526843 4742 generic.go:334] "Generic (PLEG): container finished" podID="f3b8a6f1-47db-4541-af94-5519b15d1407" containerID="052b37f81317c68a03ea64f3337d7574c49e74724e2c21557dabc01db252e9aa" exitCode=0 Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.526912 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b5xzp" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.526909 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b5xzp" event={"ID":"f3b8a6f1-47db-4541-af94-5519b15d1407","Type":"ContainerDied","Data":"052b37f81317c68a03ea64f3337d7574c49e74724e2c21557dabc01db252e9aa"} Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.527012 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b5xzp" event={"ID":"f3b8a6f1-47db-4541-af94-5519b15d1407","Type":"ContainerDied","Data":"676af860861de4fc08f05d940e72e080eae8cb9624ba2cc6d276df0354743825"} Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.542190 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-4mmhw" podStartSLOduration=1.5421676199999999 podStartE2EDuration="1.54216762s" podCreationTimestamp="2025-10-04 03:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:15:01.518630985 +0000 UTC m=+267.450968181" watchObservedRunningTime="2025-10-04 03:15:01.54216762 +0000 UTC m=+267.474504806" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.553074 4742 scope.go:117] "RemoveContainer" containerID="4a8c0ccaa12e14dfe6317a9f6c00ea1d59790e8a58ed8c1b29595dc47b51e8e3" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.556502 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-67ztt"] Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.560135 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-67ztt"] Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.570440 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-trfnh"] Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.571744 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-trfnh"] Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.583066 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lb54b"] Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.588396 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lb54b"] Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.593562 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b5xzp"] Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.593697 4742 scope.go:117] "RemoveContainer" containerID="dafdecd26505ab05deade27027f7e144160fa80a5642e21b0fcc722372082dcd" Oct 04 03:15:01 crc kubenswrapper[4742]: E1004 03:15:01.594164 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dafdecd26505ab05deade27027f7e144160fa80a5642e21b0fcc722372082dcd\": container with ID starting with dafdecd26505ab05deade27027f7e144160fa80a5642e21b0fcc722372082dcd not found: ID does not exist" containerID="dafdecd26505ab05deade27027f7e144160fa80a5642e21b0fcc722372082dcd" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.594199 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dafdecd26505ab05deade27027f7e144160fa80a5642e21b0fcc722372082dcd"} err="failed to get container status \"dafdecd26505ab05deade27027f7e144160fa80a5642e21b0fcc722372082dcd\": rpc error: code = NotFound desc = could not find container \"dafdecd26505ab05deade27027f7e144160fa80a5642e21b0fcc722372082dcd\": container with ID starting with dafdecd26505ab05deade27027f7e144160fa80a5642e21b0fcc722372082dcd not found: ID does not exist" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.594226 4742 scope.go:117] "RemoveContainer" containerID="a0b830382262c3de530a309506e7061b48f3467ae89d38d2e2ffca8b801be0ea" Oct 04 03:15:01 crc kubenswrapper[4742]: E1004 03:15:01.594661 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0b830382262c3de530a309506e7061b48f3467ae89d38d2e2ffca8b801be0ea\": container with ID starting with a0b830382262c3de530a309506e7061b48f3467ae89d38d2e2ffca8b801be0ea not found: ID does not exist" containerID="a0b830382262c3de530a309506e7061b48f3467ae89d38d2e2ffca8b801be0ea" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.594691 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0b830382262c3de530a309506e7061b48f3467ae89d38d2e2ffca8b801be0ea"} err="failed to get container status \"a0b830382262c3de530a309506e7061b48f3467ae89d38d2e2ffca8b801be0ea\": rpc error: code = NotFound desc = could not find container \"a0b830382262c3de530a309506e7061b48f3467ae89d38d2e2ffca8b801be0ea\": container with ID starting with a0b830382262c3de530a309506e7061b48f3467ae89d38d2e2ffca8b801be0ea not found: ID does not exist" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.594709 4742 scope.go:117] "RemoveContainer" containerID="4a8c0ccaa12e14dfe6317a9f6c00ea1d59790e8a58ed8c1b29595dc47b51e8e3" Oct 04 03:15:01 crc kubenswrapper[4742]: E1004 03:15:01.595029 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a8c0ccaa12e14dfe6317a9f6c00ea1d59790e8a58ed8c1b29595dc47b51e8e3\": container with ID starting with 4a8c0ccaa12e14dfe6317a9f6c00ea1d59790e8a58ed8c1b29595dc47b51e8e3 not found: ID does not exist" containerID="4a8c0ccaa12e14dfe6317a9f6c00ea1d59790e8a58ed8c1b29595dc47b51e8e3" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.595073 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a8c0ccaa12e14dfe6317a9f6c00ea1d59790e8a58ed8c1b29595dc47b51e8e3"} err="failed to get container status \"4a8c0ccaa12e14dfe6317a9f6c00ea1d59790e8a58ed8c1b29595dc47b51e8e3\": rpc error: code = NotFound desc = could not find container \"4a8c0ccaa12e14dfe6317a9f6c00ea1d59790e8a58ed8c1b29595dc47b51e8e3\": container with ID starting with 4a8c0ccaa12e14dfe6317a9f6c00ea1d59790e8a58ed8c1b29595dc47b51e8e3 not found: ID does not exist" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.595102 4742 scope.go:117] "RemoveContainer" containerID="4acb3da1f52b48c905c1821afc3f6102d530d39dfc73cc8b87ba3ef724b85dff" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.596804 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-b5xzp"] Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.603971 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tfc7g"] Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.608075 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tfc7g"] Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.610003 4742 scope.go:117] "RemoveContainer" containerID="5dde8a9c7c483d0a345f79d13dc330492b90fd476cf879785279928d4e40982a" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.625735 4742 scope.go:117] "RemoveContainer" containerID="c1283f46e50e62bf8c7be17cc716c1b39067f93311f464b7dcc25720c0655e97" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.640691 4742 scope.go:117] "RemoveContainer" containerID="4acb3da1f52b48c905c1821afc3f6102d530d39dfc73cc8b87ba3ef724b85dff" Oct 04 03:15:01 crc kubenswrapper[4742]: E1004 03:15:01.641177 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4acb3da1f52b48c905c1821afc3f6102d530d39dfc73cc8b87ba3ef724b85dff\": container with ID starting with 4acb3da1f52b48c905c1821afc3f6102d530d39dfc73cc8b87ba3ef724b85dff not found: ID does not exist" containerID="4acb3da1f52b48c905c1821afc3f6102d530d39dfc73cc8b87ba3ef724b85dff" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.641210 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4acb3da1f52b48c905c1821afc3f6102d530d39dfc73cc8b87ba3ef724b85dff"} err="failed to get container status \"4acb3da1f52b48c905c1821afc3f6102d530d39dfc73cc8b87ba3ef724b85dff\": rpc error: code = NotFound desc = could not find container \"4acb3da1f52b48c905c1821afc3f6102d530d39dfc73cc8b87ba3ef724b85dff\": container with ID starting with 4acb3da1f52b48c905c1821afc3f6102d530d39dfc73cc8b87ba3ef724b85dff not found: ID does not exist" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.641240 4742 scope.go:117] "RemoveContainer" containerID="5dde8a9c7c483d0a345f79d13dc330492b90fd476cf879785279928d4e40982a" Oct 04 03:15:01 crc kubenswrapper[4742]: E1004 03:15:01.641425 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dde8a9c7c483d0a345f79d13dc330492b90fd476cf879785279928d4e40982a\": container with ID starting with 5dde8a9c7c483d0a345f79d13dc330492b90fd476cf879785279928d4e40982a not found: ID does not exist" containerID="5dde8a9c7c483d0a345f79d13dc330492b90fd476cf879785279928d4e40982a" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.641446 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dde8a9c7c483d0a345f79d13dc330492b90fd476cf879785279928d4e40982a"} err="failed to get container status \"5dde8a9c7c483d0a345f79d13dc330492b90fd476cf879785279928d4e40982a\": rpc error: code = NotFound desc = could not find container \"5dde8a9c7c483d0a345f79d13dc330492b90fd476cf879785279928d4e40982a\": container with ID starting with 5dde8a9c7c483d0a345f79d13dc330492b90fd476cf879785279928d4e40982a not found: ID does not exist" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.641459 4742 scope.go:117] "RemoveContainer" containerID="c1283f46e50e62bf8c7be17cc716c1b39067f93311f464b7dcc25720c0655e97" Oct 04 03:15:01 crc kubenswrapper[4742]: E1004 03:15:01.641618 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1283f46e50e62bf8c7be17cc716c1b39067f93311f464b7dcc25720c0655e97\": container with ID starting with c1283f46e50e62bf8c7be17cc716c1b39067f93311f464b7dcc25720c0655e97 not found: ID does not exist" containerID="c1283f46e50e62bf8c7be17cc716c1b39067f93311f464b7dcc25720c0655e97" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.641637 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1283f46e50e62bf8c7be17cc716c1b39067f93311f464b7dcc25720c0655e97"} err="failed to get container status \"c1283f46e50e62bf8c7be17cc716c1b39067f93311f464b7dcc25720c0655e97\": rpc error: code = NotFound desc = could not find container \"c1283f46e50e62bf8c7be17cc716c1b39067f93311f464b7dcc25720c0655e97\": container with ID starting with c1283f46e50e62bf8c7be17cc716c1b39067f93311f464b7dcc25720c0655e97 not found: ID does not exist" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.641648 4742 scope.go:117] "RemoveContainer" containerID="abaa6ec10825e7beaa1a4d045a9d20823ba519919ca5549d131d09c1fb50dee6" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.660436 4742 scope.go:117] "RemoveContainer" containerID="75886b99c603a77979482570e4ab012ad56aca7cecc570aa77442da750ecef02" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.686070 4742 scope.go:117] "RemoveContainer" containerID="1d902d8de38573af3480eb1b68d8d6853cb39f8197645dba851d1fecada4d514" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.700016 4742 scope.go:117] "RemoveContainer" containerID="abaa6ec10825e7beaa1a4d045a9d20823ba519919ca5549d131d09c1fb50dee6" Oct 04 03:15:01 crc kubenswrapper[4742]: E1004 03:15:01.700467 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abaa6ec10825e7beaa1a4d045a9d20823ba519919ca5549d131d09c1fb50dee6\": container with ID starting with abaa6ec10825e7beaa1a4d045a9d20823ba519919ca5549d131d09c1fb50dee6 not found: ID does not exist" containerID="abaa6ec10825e7beaa1a4d045a9d20823ba519919ca5549d131d09c1fb50dee6" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.700496 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abaa6ec10825e7beaa1a4d045a9d20823ba519919ca5549d131d09c1fb50dee6"} err="failed to get container status \"abaa6ec10825e7beaa1a4d045a9d20823ba519919ca5549d131d09c1fb50dee6\": rpc error: code = NotFound desc = could not find container \"abaa6ec10825e7beaa1a4d045a9d20823ba519919ca5549d131d09c1fb50dee6\": container with ID starting with abaa6ec10825e7beaa1a4d045a9d20823ba519919ca5549d131d09c1fb50dee6 not found: ID does not exist" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.700523 4742 scope.go:117] "RemoveContainer" containerID="75886b99c603a77979482570e4ab012ad56aca7cecc570aa77442da750ecef02" Oct 04 03:15:01 crc kubenswrapper[4742]: E1004 03:15:01.700969 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75886b99c603a77979482570e4ab012ad56aca7cecc570aa77442da750ecef02\": container with ID starting with 75886b99c603a77979482570e4ab012ad56aca7cecc570aa77442da750ecef02 not found: ID does not exist" containerID="75886b99c603a77979482570e4ab012ad56aca7cecc570aa77442da750ecef02" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.701091 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75886b99c603a77979482570e4ab012ad56aca7cecc570aa77442da750ecef02"} err="failed to get container status \"75886b99c603a77979482570e4ab012ad56aca7cecc570aa77442da750ecef02\": rpc error: code = NotFound desc = could not find container \"75886b99c603a77979482570e4ab012ad56aca7cecc570aa77442da750ecef02\": container with ID starting with 75886b99c603a77979482570e4ab012ad56aca7cecc570aa77442da750ecef02 not found: ID does not exist" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.701193 4742 scope.go:117] "RemoveContainer" containerID="1d902d8de38573af3480eb1b68d8d6853cb39f8197645dba851d1fecada4d514" Oct 04 03:15:01 crc kubenswrapper[4742]: E1004 03:15:01.701626 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d902d8de38573af3480eb1b68d8d6853cb39f8197645dba851d1fecada4d514\": container with ID starting with 1d902d8de38573af3480eb1b68d8d6853cb39f8197645dba851d1fecada4d514 not found: ID does not exist" containerID="1d902d8de38573af3480eb1b68d8d6853cb39f8197645dba851d1fecada4d514" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.701747 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d902d8de38573af3480eb1b68d8d6853cb39f8197645dba851d1fecada4d514"} err="failed to get container status \"1d902d8de38573af3480eb1b68d8d6853cb39f8197645dba851d1fecada4d514\": rpc error: code = NotFound desc = could not find container \"1d902d8de38573af3480eb1b68d8d6853cb39f8197645dba851d1fecada4d514\": container with ID starting with 1d902d8de38573af3480eb1b68d8d6853cb39f8197645dba851d1fecada4d514 not found: ID does not exist" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.701836 4742 scope.go:117] "RemoveContainer" containerID="9f9c46a4ab5b18d70d78e1592b5d352927e2103c6d485583eb027a29fe3d3fa5" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.715129 4742 scope.go:117] "RemoveContainer" containerID="9f9c46a4ab5b18d70d78e1592b5d352927e2103c6d485583eb027a29fe3d3fa5" Oct 04 03:15:01 crc kubenswrapper[4742]: E1004 03:15:01.715816 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f9c46a4ab5b18d70d78e1592b5d352927e2103c6d485583eb027a29fe3d3fa5\": container with ID starting with 9f9c46a4ab5b18d70d78e1592b5d352927e2103c6d485583eb027a29fe3d3fa5 not found: ID does not exist" containerID="9f9c46a4ab5b18d70d78e1592b5d352927e2103c6d485583eb027a29fe3d3fa5" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.715844 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f9c46a4ab5b18d70d78e1592b5d352927e2103c6d485583eb027a29fe3d3fa5"} err="failed to get container status \"9f9c46a4ab5b18d70d78e1592b5d352927e2103c6d485583eb027a29fe3d3fa5\": rpc error: code = NotFound desc = could not find container \"9f9c46a4ab5b18d70d78e1592b5d352927e2103c6d485583eb027a29fe3d3fa5\": container with ID starting with 9f9c46a4ab5b18d70d78e1592b5d352927e2103c6d485583eb027a29fe3d3fa5 not found: ID does not exist" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.715868 4742 scope.go:117] "RemoveContainer" containerID="052b37f81317c68a03ea64f3337d7574c49e74724e2c21557dabc01db252e9aa" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.731155 4742 scope.go:117] "RemoveContainer" containerID="1e188d2c4d79496549160f22be90e4a093df16f4663a5910ba90d607e2ea6f86" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.750311 4742 scope.go:117] "RemoveContainer" containerID="452d6d3a8866d451b5f40bd252e13295fd55e9a672b02d533f288bd323d389eb" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.763366 4742 scope.go:117] "RemoveContainer" containerID="052b37f81317c68a03ea64f3337d7574c49e74724e2c21557dabc01db252e9aa" Oct 04 03:15:01 crc kubenswrapper[4742]: E1004 03:15:01.763894 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"052b37f81317c68a03ea64f3337d7574c49e74724e2c21557dabc01db252e9aa\": container with ID starting with 052b37f81317c68a03ea64f3337d7574c49e74724e2c21557dabc01db252e9aa not found: ID does not exist" containerID="052b37f81317c68a03ea64f3337d7574c49e74724e2c21557dabc01db252e9aa" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.763929 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"052b37f81317c68a03ea64f3337d7574c49e74724e2c21557dabc01db252e9aa"} err="failed to get container status \"052b37f81317c68a03ea64f3337d7574c49e74724e2c21557dabc01db252e9aa\": rpc error: code = NotFound desc = could not find container \"052b37f81317c68a03ea64f3337d7574c49e74724e2c21557dabc01db252e9aa\": container with ID starting with 052b37f81317c68a03ea64f3337d7574c49e74724e2c21557dabc01db252e9aa not found: ID does not exist" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.763962 4742 scope.go:117] "RemoveContainer" containerID="1e188d2c4d79496549160f22be90e4a093df16f4663a5910ba90d607e2ea6f86" Oct 04 03:15:01 crc kubenswrapper[4742]: E1004 03:15:01.764297 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e188d2c4d79496549160f22be90e4a093df16f4663a5910ba90d607e2ea6f86\": container with ID starting with 1e188d2c4d79496549160f22be90e4a093df16f4663a5910ba90d607e2ea6f86 not found: ID does not exist" containerID="1e188d2c4d79496549160f22be90e4a093df16f4663a5910ba90d607e2ea6f86" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.764349 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e188d2c4d79496549160f22be90e4a093df16f4663a5910ba90d607e2ea6f86"} err="failed to get container status \"1e188d2c4d79496549160f22be90e4a093df16f4663a5910ba90d607e2ea6f86\": rpc error: code = NotFound desc = could not find container \"1e188d2c4d79496549160f22be90e4a093df16f4663a5910ba90d607e2ea6f86\": container with ID starting with 1e188d2c4d79496549160f22be90e4a093df16f4663a5910ba90d607e2ea6f86 not found: ID does not exist" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.764420 4742 scope.go:117] "RemoveContainer" containerID="452d6d3a8866d451b5f40bd252e13295fd55e9a672b02d533f288bd323d389eb" Oct 04 03:15:01 crc kubenswrapper[4742]: E1004 03:15:01.764769 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"452d6d3a8866d451b5f40bd252e13295fd55e9a672b02d533f288bd323d389eb\": container with ID starting with 452d6d3a8866d451b5f40bd252e13295fd55e9a672b02d533f288bd323d389eb not found: ID does not exist" containerID="452d6d3a8866d451b5f40bd252e13295fd55e9a672b02d533f288bd323d389eb" Oct 04 03:15:01 crc kubenswrapper[4742]: I1004 03:15:01.764809 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"452d6d3a8866d451b5f40bd252e13295fd55e9a672b02d533f288bd323d389eb"} err="failed to get container status \"452d6d3a8866d451b5f40bd252e13295fd55e9a672b02d533f288bd323d389eb\": rpc error: code = NotFound desc = could not find container \"452d6d3a8866d451b5f40bd252e13295fd55e9a672b02d533f288bd323d389eb\": container with ID starting with 452d6d3a8866d451b5f40bd252e13295fd55e9a672b02d533f288bd323d389eb not found: ID does not exist" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.542632 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-4mmhw" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.830177 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-c29rn"] Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.830605 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325795-vhpcq" Oct 04 03:15:02 crc kubenswrapper[4742]: E1004 03:15:02.830948 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2788ab92-5c72-4f27-888e-ef695e4cf6cb" containerName="registry-server" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.830962 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="2788ab92-5c72-4f27-888e-ef695e4cf6cb" containerName="registry-server" Oct 04 03:15:02 crc kubenswrapper[4742]: E1004 03:15:02.830976 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3b8a6f1-47db-4541-af94-5519b15d1407" containerName="extract-utilities" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.830984 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3b8a6f1-47db-4541-af94-5519b15d1407" containerName="extract-utilities" Oct 04 03:15:02 crc kubenswrapper[4742]: E1004 03:15:02.830996 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3fa495f-e19d-4658-a335-1e49cb14a721" containerName="extract-utilities" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.831003 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3fa495f-e19d-4658-a335-1e49cb14a721" containerName="extract-utilities" Oct 04 03:15:02 crc kubenswrapper[4742]: E1004 03:15:02.831015 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c210acab-8945-4f9d-a8fe-47ae7d78a064" containerName="registry-server" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.831023 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="c210acab-8945-4f9d-a8fe-47ae7d78a064" containerName="registry-server" Oct 04 03:15:02 crc kubenswrapper[4742]: E1004 03:15:02.831036 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71e52223-cb12-4aae-8dcf-7418f39073c7" containerName="collect-profiles" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.831043 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="71e52223-cb12-4aae-8dcf-7418f39073c7" containerName="collect-profiles" Oct 04 03:15:02 crc kubenswrapper[4742]: E1004 03:15:02.831051 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3b8a6f1-47db-4541-af94-5519b15d1407" containerName="extract-content" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.831058 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3b8a6f1-47db-4541-af94-5519b15d1407" containerName="extract-content" Oct 04 03:15:02 crc kubenswrapper[4742]: E1004 03:15:02.831067 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c210acab-8945-4f9d-a8fe-47ae7d78a064" containerName="extract-content" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.831074 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="c210acab-8945-4f9d-a8fe-47ae7d78a064" containerName="extract-content" Oct 04 03:15:02 crc kubenswrapper[4742]: E1004 03:15:02.831082 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2788ab92-5c72-4f27-888e-ef695e4cf6cb" containerName="extract-content" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.831087 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="2788ab92-5c72-4f27-888e-ef695e4cf6cb" containerName="extract-content" Oct 04 03:15:02 crc kubenswrapper[4742]: E1004 03:15:02.831094 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2788ab92-5c72-4f27-888e-ef695e4cf6cb" containerName="extract-utilities" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.831100 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="2788ab92-5c72-4f27-888e-ef695e4cf6cb" containerName="extract-utilities" Oct 04 03:15:02 crc kubenswrapper[4742]: E1004 03:15:02.831108 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3fa495f-e19d-4658-a335-1e49cb14a721" containerName="extract-content" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.831113 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3fa495f-e19d-4658-a335-1e49cb14a721" containerName="extract-content" Oct 04 03:15:02 crc kubenswrapper[4742]: E1004 03:15:02.831123 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3b8a6f1-47db-4541-af94-5519b15d1407" containerName="registry-server" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.831128 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3b8a6f1-47db-4541-af94-5519b15d1407" containerName="registry-server" Oct 04 03:15:02 crc kubenswrapper[4742]: E1004 03:15:02.831135 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3fa495f-e19d-4658-a335-1e49cb14a721" containerName="registry-server" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.831140 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3fa495f-e19d-4658-a335-1e49cb14a721" containerName="registry-server" Oct 04 03:15:02 crc kubenswrapper[4742]: E1004 03:15:02.831147 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c210acab-8945-4f9d-a8fe-47ae7d78a064" containerName="extract-utilities" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.831153 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="c210acab-8945-4f9d-a8fe-47ae7d78a064" containerName="extract-utilities" Oct 04 03:15:02 crc kubenswrapper[4742]: E1004 03:15:02.831162 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8b41c08-4e89-4f08-8d45-256c306fbea0" containerName="marketplace-operator" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.831169 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8b41c08-4e89-4f08-8d45-256c306fbea0" containerName="marketplace-operator" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.831261 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="71e52223-cb12-4aae-8dcf-7418f39073c7" containerName="collect-profiles" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.831285 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="c210acab-8945-4f9d-a8fe-47ae7d78a064" containerName="registry-server" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.831294 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="2788ab92-5c72-4f27-888e-ef695e4cf6cb" containerName="registry-server" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.831301 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8b41c08-4e89-4f08-8d45-256c306fbea0" containerName="marketplace-operator" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.831312 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3fa495f-e19d-4658-a335-1e49cb14a721" containerName="registry-server" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.831322 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3b8a6f1-47db-4541-af94-5519b15d1407" containerName="registry-server" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.832012 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c29rn" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.836744 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.843372 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c29rn"] Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.865344 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2788ab92-5c72-4f27-888e-ef695e4cf6cb" path="/var/lib/kubelet/pods/2788ab92-5c72-4f27-888e-ef695e4cf6cb/volumes" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.865956 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8b41c08-4e89-4f08-8d45-256c306fbea0" path="/var/lib/kubelet/pods/a8b41c08-4e89-4f08-8d45-256c306fbea0/volumes" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.866406 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c210acab-8945-4f9d-a8fe-47ae7d78a064" path="/var/lib/kubelet/pods/c210acab-8945-4f9d-a8fe-47ae7d78a064/volumes" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.867628 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3fa495f-e19d-4658-a335-1e49cb14a721" path="/var/lib/kubelet/pods/c3fa495f-e19d-4658-a335-1e49cb14a721/volumes" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.868360 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3b8a6f1-47db-4541-af94-5519b15d1407" path="/var/lib/kubelet/pods/f3b8a6f1-47db-4541-af94-5519b15d1407/volumes" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.885406 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ac418c2-2977-4273-897b-282809b825d0-catalog-content\") pod \"redhat-marketplace-c29rn\" (UID: \"1ac418c2-2977-4273-897b-282809b825d0\") " pod="openshift-marketplace/redhat-marketplace-c29rn" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.885462 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ac418c2-2977-4273-897b-282809b825d0-utilities\") pod \"redhat-marketplace-c29rn\" (UID: \"1ac418c2-2977-4273-897b-282809b825d0\") " pod="openshift-marketplace/redhat-marketplace-c29rn" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.885504 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mccg2\" (UniqueName: \"kubernetes.io/projected/1ac418c2-2977-4273-897b-282809b825d0-kube-api-access-mccg2\") pod \"redhat-marketplace-c29rn\" (UID: \"1ac418c2-2977-4273-897b-282809b825d0\") " pod="openshift-marketplace/redhat-marketplace-c29rn" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.985979 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/71e52223-cb12-4aae-8dcf-7418f39073c7-config-volume\") pod \"71e52223-cb12-4aae-8dcf-7418f39073c7\" (UID: \"71e52223-cb12-4aae-8dcf-7418f39073c7\") " Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.986071 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hp5q6\" (UniqueName: \"kubernetes.io/projected/71e52223-cb12-4aae-8dcf-7418f39073c7-kube-api-access-hp5q6\") pod \"71e52223-cb12-4aae-8dcf-7418f39073c7\" (UID: \"71e52223-cb12-4aae-8dcf-7418f39073c7\") " Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.986139 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/71e52223-cb12-4aae-8dcf-7418f39073c7-secret-volume\") pod \"71e52223-cb12-4aae-8dcf-7418f39073c7\" (UID: \"71e52223-cb12-4aae-8dcf-7418f39073c7\") " Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.986289 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mccg2\" (UniqueName: \"kubernetes.io/projected/1ac418c2-2977-4273-897b-282809b825d0-kube-api-access-mccg2\") pod \"redhat-marketplace-c29rn\" (UID: \"1ac418c2-2977-4273-897b-282809b825d0\") " pod="openshift-marketplace/redhat-marketplace-c29rn" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.986338 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ac418c2-2977-4273-897b-282809b825d0-catalog-content\") pod \"redhat-marketplace-c29rn\" (UID: \"1ac418c2-2977-4273-897b-282809b825d0\") " pod="openshift-marketplace/redhat-marketplace-c29rn" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.986367 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ac418c2-2977-4273-897b-282809b825d0-utilities\") pod \"redhat-marketplace-c29rn\" (UID: \"1ac418c2-2977-4273-897b-282809b825d0\") " pod="openshift-marketplace/redhat-marketplace-c29rn" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.986780 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ac418c2-2977-4273-897b-282809b825d0-utilities\") pod \"redhat-marketplace-c29rn\" (UID: \"1ac418c2-2977-4273-897b-282809b825d0\") " pod="openshift-marketplace/redhat-marketplace-c29rn" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.986942 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ac418c2-2977-4273-897b-282809b825d0-catalog-content\") pod \"redhat-marketplace-c29rn\" (UID: \"1ac418c2-2977-4273-897b-282809b825d0\") " pod="openshift-marketplace/redhat-marketplace-c29rn" Oct 04 03:15:02 crc kubenswrapper[4742]: I1004 03:15:02.987221 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71e52223-cb12-4aae-8dcf-7418f39073c7-config-volume" (OuterVolumeSpecName: "config-volume") pod "71e52223-cb12-4aae-8dcf-7418f39073c7" (UID: "71e52223-cb12-4aae-8dcf-7418f39073c7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:15:03 crc kubenswrapper[4742]: I1004 03:15:03.001863 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mccg2\" (UniqueName: \"kubernetes.io/projected/1ac418c2-2977-4273-897b-282809b825d0-kube-api-access-mccg2\") pod \"redhat-marketplace-c29rn\" (UID: \"1ac418c2-2977-4273-897b-282809b825d0\") " pod="openshift-marketplace/redhat-marketplace-c29rn" Oct 04 03:15:03 crc kubenswrapper[4742]: I1004 03:15:03.002999 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71e52223-cb12-4aae-8dcf-7418f39073c7-kube-api-access-hp5q6" (OuterVolumeSpecName: "kube-api-access-hp5q6") pod "71e52223-cb12-4aae-8dcf-7418f39073c7" (UID: "71e52223-cb12-4aae-8dcf-7418f39073c7"). InnerVolumeSpecName "kube-api-access-hp5q6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:15:03 crc kubenswrapper[4742]: I1004 03:15:03.003373 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71e52223-cb12-4aae-8dcf-7418f39073c7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "71e52223-cb12-4aae-8dcf-7418f39073c7" (UID: "71e52223-cb12-4aae-8dcf-7418f39073c7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:15:03 crc kubenswrapper[4742]: I1004 03:15:03.034237 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9ztnx"] Oct 04 03:15:03 crc kubenswrapper[4742]: I1004 03:15:03.042390 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9ztnx" Oct 04 03:15:03 crc kubenswrapper[4742]: I1004 03:15:03.044095 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 04 03:15:03 crc kubenswrapper[4742]: I1004 03:15:03.048304 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9ztnx"] Oct 04 03:15:03 crc kubenswrapper[4742]: I1004 03:15:03.088077 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4f312f8-6a5a-41da-9de0-0e67ea5b3344-utilities\") pod \"redhat-operators-9ztnx\" (UID: \"f4f312f8-6a5a-41da-9de0-0e67ea5b3344\") " pod="openshift-marketplace/redhat-operators-9ztnx" Oct 04 03:15:03 crc kubenswrapper[4742]: I1004 03:15:03.088160 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gf55\" (UniqueName: \"kubernetes.io/projected/f4f312f8-6a5a-41da-9de0-0e67ea5b3344-kube-api-access-9gf55\") pod \"redhat-operators-9ztnx\" (UID: \"f4f312f8-6a5a-41da-9de0-0e67ea5b3344\") " pod="openshift-marketplace/redhat-operators-9ztnx" Oct 04 03:15:03 crc kubenswrapper[4742]: I1004 03:15:03.088239 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4f312f8-6a5a-41da-9de0-0e67ea5b3344-catalog-content\") pod \"redhat-operators-9ztnx\" (UID: \"f4f312f8-6a5a-41da-9de0-0e67ea5b3344\") " pod="openshift-marketplace/redhat-operators-9ztnx" Oct 04 03:15:03 crc kubenswrapper[4742]: I1004 03:15:03.088322 4742 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/71e52223-cb12-4aae-8dcf-7418f39073c7-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:03 crc kubenswrapper[4742]: I1004 03:15:03.088341 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hp5q6\" (UniqueName: \"kubernetes.io/projected/71e52223-cb12-4aae-8dcf-7418f39073c7-kube-api-access-hp5q6\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:03 crc kubenswrapper[4742]: I1004 03:15:03.088356 4742 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/71e52223-cb12-4aae-8dcf-7418f39073c7-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 03:15:03 crc kubenswrapper[4742]: I1004 03:15:03.148753 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c29rn" Oct 04 03:15:03 crc kubenswrapper[4742]: I1004 03:15:03.189199 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gf55\" (UniqueName: \"kubernetes.io/projected/f4f312f8-6a5a-41da-9de0-0e67ea5b3344-kube-api-access-9gf55\") pod \"redhat-operators-9ztnx\" (UID: \"f4f312f8-6a5a-41da-9de0-0e67ea5b3344\") " pod="openshift-marketplace/redhat-operators-9ztnx" Oct 04 03:15:03 crc kubenswrapper[4742]: I1004 03:15:03.189311 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4f312f8-6a5a-41da-9de0-0e67ea5b3344-catalog-content\") pod \"redhat-operators-9ztnx\" (UID: \"f4f312f8-6a5a-41da-9de0-0e67ea5b3344\") " pod="openshift-marketplace/redhat-operators-9ztnx" Oct 04 03:15:03 crc kubenswrapper[4742]: I1004 03:15:03.189366 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4f312f8-6a5a-41da-9de0-0e67ea5b3344-utilities\") pod \"redhat-operators-9ztnx\" (UID: \"f4f312f8-6a5a-41da-9de0-0e67ea5b3344\") " pod="openshift-marketplace/redhat-operators-9ztnx" Oct 04 03:15:03 crc kubenswrapper[4742]: I1004 03:15:03.189895 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4f312f8-6a5a-41da-9de0-0e67ea5b3344-utilities\") pod \"redhat-operators-9ztnx\" (UID: \"f4f312f8-6a5a-41da-9de0-0e67ea5b3344\") " pod="openshift-marketplace/redhat-operators-9ztnx" Oct 04 03:15:03 crc kubenswrapper[4742]: I1004 03:15:03.189931 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4f312f8-6a5a-41da-9de0-0e67ea5b3344-catalog-content\") pod \"redhat-operators-9ztnx\" (UID: \"f4f312f8-6a5a-41da-9de0-0e67ea5b3344\") " pod="openshift-marketplace/redhat-operators-9ztnx" Oct 04 03:15:03 crc kubenswrapper[4742]: I1004 03:15:03.204959 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gf55\" (UniqueName: \"kubernetes.io/projected/f4f312f8-6a5a-41da-9de0-0e67ea5b3344-kube-api-access-9gf55\") pod \"redhat-operators-9ztnx\" (UID: \"f4f312f8-6a5a-41da-9de0-0e67ea5b3344\") " pod="openshift-marketplace/redhat-operators-9ztnx" Oct 04 03:15:03 crc kubenswrapper[4742]: I1004 03:15:03.336511 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c29rn"] Oct 04 03:15:03 crc kubenswrapper[4742]: W1004 03:15:03.345827 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ac418c2_2977_4273_897b_282809b825d0.slice/crio-56beb470723c8578c98112eda97b2b383afb446b251417e64cd07eb6a039b421 WatchSource:0}: Error finding container 56beb470723c8578c98112eda97b2b383afb446b251417e64cd07eb6a039b421: Status 404 returned error can't find the container with id 56beb470723c8578c98112eda97b2b383afb446b251417e64cd07eb6a039b421 Oct 04 03:15:03 crc kubenswrapper[4742]: I1004 03:15:03.357393 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9ztnx" Oct 04 03:15:03 crc kubenswrapper[4742]: I1004 03:15:03.530294 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9ztnx"] Oct 04 03:15:03 crc kubenswrapper[4742]: I1004 03:15:03.545814 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325795-vhpcq" event={"ID":"71e52223-cb12-4aae-8dcf-7418f39073c7","Type":"ContainerDied","Data":"a0071d46424c0640c2bec58fa70821a3b772e95d24849d8bd8f93f335138fc9f"} Oct 04 03:15:03 crc kubenswrapper[4742]: I1004 03:15:03.545870 4742 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0071d46424c0640c2bec58fa70821a3b772e95d24849d8bd8f93f335138fc9f" Oct 04 03:15:03 crc kubenswrapper[4742]: I1004 03:15:03.545838 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325795-vhpcq" Oct 04 03:15:03 crc kubenswrapper[4742]: I1004 03:15:03.560473 4742 generic.go:334] "Generic (PLEG): container finished" podID="1ac418c2-2977-4273-897b-282809b825d0" containerID="c5f80eb2216638c317946109e36f6ed150d94438002500869216ed2f47239eaf" exitCode=0 Oct 04 03:15:03 crc kubenswrapper[4742]: I1004 03:15:03.560588 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c29rn" event={"ID":"1ac418c2-2977-4273-897b-282809b825d0","Type":"ContainerDied","Data":"c5f80eb2216638c317946109e36f6ed150d94438002500869216ed2f47239eaf"} Oct 04 03:15:03 crc kubenswrapper[4742]: I1004 03:15:03.560621 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c29rn" event={"ID":"1ac418c2-2977-4273-897b-282809b825d0","Type":"ContainerStarted","Data":"56beb470723c8578c98112eda97b2b383afb446b251417e64cd07eb6a039b421"} Oct 04 03:15:04 crc kubenswrapper[4742]: I1004 03:15:04.568617 4742 generic.go:334] "Generic (PLEG): container finished" podID="1ac418c2-2977-4273-897b-282809b825d0" containerID="65fad14dd4b8a8583450dd3757aa73934b498a8322d9fc204ce7b7eb63956234" exitCode=0 Oct 04 03:15:04 crc kubenswrapper[4742]: I1004 03:15:04.568715 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c29rn" event={"ID":"1ac418c2-2977-4273-897b-282809b825d0","Type":"ContainerDied","Data":"65fad14dd4b8a8583450dd3757aa73934b498a8322d9fc204ce7b7eb63956234"} Oct 04 03:15:04 crc kubenswrapper[4742]: I1004 03:15:04.573776 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9ztnx" event={"ID":"f4f312f8-6a5a-41da-9de0-0e67ea5b3344","Type":"ContainerDied","Data":"c4913b00ace9b6f33ce799752abfaaeebffb24a62a2f78093aa74e10833d9007"} Oct 04 03:15:04 crc kubenswrapper[4742]: I1004 03:15:04.571914 4742 generic.go:334] "Generic (PLEG): container finished" podID="f4f312f8-6a5a-41da-9de0-0e67ea5b3344" containerID="c4913b00ace9b6f33ce799752abfaaeebffb24a62a2f78093aa74e10833d9007" exitCode=0 Oct 04 03:15:04 crc kubenswrapper[4742]: I1004 03:15:04.574047 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9ztnx" event={"ID":"f4f312f8-6a5a-41da-9de0-0e67ea5b3344","Type":"ContainerStarted","Data":"f819a87a14bfcf0281fac14a0ada848a99059a53e36076fd55cee91813480a29"} Oct 04 03:15:05 crc kubenswrapper[4742]: I1004 03:15:05.235063 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-l5vff"] Oct 04 03:15:05 crc kubenswrapper[4742]: I1004 03:15:05.237074 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l5vff" Oct 04 03:15:05 crc kubenswrapper[4742]: I1004 03:15:05.239877 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 04 03:15:05 crc kubenswrapper[4742]: I1004 03:15:05.245526 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l5vff"] Oct 04 03:15:05 crc kubenswrapper[4742]: I1004 03:15:05.415863 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e50888f-9b21-423e-a450-66f235ad93c8-catalog-content\") pod \"certified-operators-l5vff\" (UID: \"3e50888f-9b21-423e-a450-66f235ad93c8\") " pod="openshift-marketplace/certified-operators-l5vff" Oct 04 03:15:05 crc kubenswrapper[4742]: I1004 03:15:05.415961 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkfc9\" (UniqueName: \"kubernetes.io/projected/3e50888f-9b21-423e-a450-66f235ad93c8-kube-api-access-bkfc9\") pod \"certified-operators-l5vff\" (UID: \"3e50888f-9b21-423e-a450-66f235ad93c8\") " pod="openshift-marketplace/certified-operators-l5vff" Oct 04 03:15:05 crc kubenswrapper[4742]: I1004 03:15:05.416020 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e50888f-9b21-423e-a450-66f235ad93c8-utilities\") pod \"certified-operators-l5vff\" (UID: \"3e50888f-9b21-423e-a450-66f235ad93c8\") " pod="openshift-marketplace/certified-operators-l5vff" Oct 04 03:15:05 crc kubenswrapper[4742]: I1004 03:15:05.433841 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-t2zpx"] Oct 04 03:15:05 crc kubenswrapper[4742]: I1004 03:15:05.435043 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t2zpx" Oct 04 03:15:05 crc kubenswrapper[4742]: I1004 03:15:05.436968 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 04 03:15:05 crc kubenswrapper[4742]: I1004 03:15:05.445732 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t2zpx"] Oct 04 03:15:05 crc kubenswrapper[4742]: I1004 03:15:05.516773 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkfc9\" (UniqueName: \"kubernetes.io/projected/3e50888f-9b21-423e-a450-66f235ad93c8-kube-api-access-bkfc9\") pod \"certified-operators-l5vff\" (UID: \"3e50888f-9b21-423e-a450-66f235ad93c8\") " pod="openshift-marketplace/certified-operators-l5vff" Oct 04 03:15:05 crc kubenswrapper[4742]: I1004 03:15:05.516958 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e50888f-9b21-423e-a450-66f235ad93c8-utilities\") pod \"certified-operators-l5vff\" (UID: \"3e50888f-9b21-423e-a450-66f235ad93c8\") " pod="openshift-marketplace/certified-operators-l5vff" Oct 04 03:15:05 crc kubenswrapper[4742]: I1004 03:15:05.517043 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e50888f-9b21-423e-a450-66f235ad93c8-catalog-content\") pod \"certified-operators-l5vff\" (UID: \"3e50888f-9b21-423e-a450-66f235ad93c8\") " pod="openshift-marketplace/certified-operators-l5vff" Oct 04 03:15:05 crc kubenswrapper[4742]: I1004 03:15:05.517483 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e50888f-9b21-423e-a450-66f235ad93c8-utilities\") pod \"certified-operators-l5vff\" (UID: \"3e50888f-9b21-423e-a450-66f235ad93c8\") " pod="openshift-marketplace/certified-operators-l5vff" Oct 04 03:15:05 crc kubenswrapper[4742]: I1004 03:15:05.517553 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e50888f-9b21-423e-a450-66f235ad93c8-catalog-content\") pod \"certified-operators-l5vff\" (UID: \"3e50888f-9b21-423e-a450-66f235ad93c8\") " pod="openshift-marketplace/certified-operators-l5vff" Oct 04 03:15:05 crc kubenswrapper[4742]: I1004 03:15:05.543305 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkfc9\" (UniqueName: \"kubernetes.io/projected/3e50888f-9b21-423e-a450-66f235ad93c8-kube-api-access-bkfc9\") pod \"certified-operators-l5vff\" (UID: \"3e50888f-9b21-423e-a450-66f235ad93c8\") " pod="openshift-marketplace/certified-operators-l5vff" Oct 04 03:15:05 crc kubenswrapper[4742]: I1004 03:15:05.580306 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c29rn" event={"ID":"1ac418c2-2977-4273-897b-282809b825d0","Type":"ContainerStarted","Data":"8c75fad8ca91b1a723ea7032f990def15e0456c946b5b45f76f56725fae9d914"} Oct 04 03:15:05 crc kubenswrapper[4742]: I1004 03:15:05.598413 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-c29rn" podStartSLOduration=1.9426627509999999 podStartE2EDuration="3.598392486s" podCreationTimestamp="2025-10-04 03:15:02 +0000 UTC" firstStartedPulling="2025-10-04 03:15:03.563680356 +0000 UTC m=+269.496017522" lastFinishedPulling="2025-10-04 03:15:05.219410081 +0000 UTC m=+271.151747257" observedRunningTime="2025-10-04 03:15:05.597897824 +0000 UTC m=+271.530235000" watchObservedRunningTime="2025-10-04 03:15:05.598392486 +0000 UTC m=+271.530729662" Oct 04 03:15:05 crc kubenswrapper[4742]: I1004 03:15:05.617834 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bde2e58-49b8-4975-aaaf-2cecd14c34d1-utilities\") pod \"community-operators-t2zpx\" (UID: \"4bde2e58-49b8-4975-aaaf-2cecd14c34d1\") " pod="openshift-marketplace/community-operators-t2zpx" Oct 04 03:15:05 crc kubenswrapper[4742]: I1004 03:15:05.617886 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bde2e58-49b8-4975-aaaf-2cecd14c34d1-catalog-content\") pod \"community-operators-t2zpx\" (UID: \"4bde2e58-49b8-4975-aaaf-2cecd14c34d1\") " pod="openshift-marketplace/community-operators-t2zpx" Oct 04 03:15:05 crc kubenswrapper[4742]: I1004 03:15:05.617954 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvb78\" (UniqueName: \"kubernetes.io/projected/4bde2e58-49b8-4975-aaaf-2cecd14c34d1-kube-api-access-fvb78\") pod \"community-operators-t2zpx\" (UID: \"4bde2e58-49b8-4975-aaaf-2cecd14c34d1\") " pod="openshift-marketplace/community-operators-t2zpx" Oct 04 03:15:05 crc kubenswrapper[4742]: I1004 03:15:05.631378 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l5vff" Oct 04 03:15:05 crc kubenswrapper[4742]: I1004 03:15:05.718876 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvb78\" (UniqueName: \"kubernetes.io/projected/4bde2e58-49b8-4975-aaaf-2cecd14c34d1-kube-api-access-fvb78\") pod \"community-operators-t2zpx\" (UID: \"4bde2e58-49b8-4975-aaaf-2cecd14c34d1\") " pod="openshift-marketplace/community-operators-t2zpx" Oct 04 03:15:05 crc kubenswrapper[4742]: I1004 03:15:05.718938 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bde2e58-49b8-4975-aaaf-2cecd14c34d1-utilities\") pod \"community-operators-t2zpx\" (UID: \"4bde2e58-49b8-4975-aaaf-2cecd14c34d1\") " pod="openshift-marketplace/community-operators-t2zpx" Oct 04 03:15:05 crc kubenswrapper[4742]: I1004 03:15:05.718963 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bde2e58-49b8-4975-aaaf-2cecd14c34d1-catalog-content\") pod \"community-operators-t2zpx\" (UID: \"4bde2e58-49b8-4975-aaaf-2cecd14c34d1\") " pod="openshift-marketplace/community-operators-t2zpx" Oct 04 03:15:05 crc kubenswrapper[4742]: I1004 03:15:05.719765 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bde2e58-49b8-4975-aaaf-2cecd14c34d1-utilities\") pod \"community-operators-t2zpx\" (UID: \"4bde2e58-49b8-4975-aaaf-2cecd14c34d1\") " pod="openshift-marketplace/community-operators-t2zpx" Oct 04 03:15:05 crc kubenswrapper[4742]: I1004 03:15:05.726078 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bde2e58-49b8-4975-aaaf-2cecd14c34d1-catalog-content\") pod \"community-operators-t2zpx\" (UID: \"4bde2e58-49b8-4975-aaaf-2cecd14c34d1\") " pod="openshift-marketplace/community-operators-t2zpx" Oct 04 03:15:05 crc kubenswrapper[4742]: I1004 03:15:05.743114 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvb78\" (UniqueName: \"kubernetes.io/projected/4bde2e58-49b8-4975-aaaf-2cecd14c34d1-kube-api-access-fvb78\") pod \"community-operators-t2zpx\" (UID: \"4bde2e58-49b8-4975-aaaf-2cecd14c34d1\") " pod="openshift-marketplace/community-operators-t2zpx" Oct 04 03:15:05 crc kubenswrapper[4742]: I1004 03:15:05.748668 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t2zpx" Oct 04 03:15:05 crc kubenswrapper[4742]: I1004 03:15:05.838404 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l5vff"] Oct 04 03:15:06 crc kubenswrapper[4742]: I1004 03:15:06.014116 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t2zpx"] Oct 04 03:15:06 crc kubenswrapper[4742]: I1004 03:15:06.587817 4742 generic.go:334] "Generic (PLEG): container finished" podID="4bde2e58-49b8-4975-aaaf-2cecd14c34d1" containerID="18673902019c1b2eea7c1b5cb998d7bc06782ef06a5273d69c0eb8a14663a2c5" exitCode=0 Oct 04 03:15:06 crc kubenswrapper[4742]: I1004 03:15:06.588199 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t2zpx" event={"ID":"4bde2e58-49b8-4975-aaaf-2cecd14c34d1","Type":"ContainerDied","Data":"18673902019c1b2eea7c1b5cb998d7bc06782ef06a5273d69c0eb8a14663a2c5"} Oct 04 03:15:06 crc kubenswrapper[4742]: I1004 03:15:06.588287 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t2zpx" event={"ID":"4bde2e58-49b8-4975-aaaf-2cecd14c34d1","Type":"ContainerStarted","Data":"98355b695e046c1115b3301f10c1a8037cc5f90604edf73dd682d60a703755ad"} Oct 04 03:15:06 crc kubenswrapper[4742]: I1004 03:15:06.592358 4742 generic.go:334] "Generic (PLEG): container finished" podID="f4f312f8-6a5a-41da-9de0-0e67ea5b3344" containerID="2edb66b2157f6cc712e328cceb6d9f0c0860f00b607c1f458e94323c4e142d74" exitCode=0 Oct 04 03:15:06 crc kubenswrapper[4742]: I1004 03:15:06.592472 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9ztnx" event={"ID":"f4f312f8-6a5a-41da-9de0-0e67ea5b3344","Type":"ContainerDied","Data":"2edb66b2157f6cc712e328cceb6d9f0c0860f00b607c1f458e94323c4e142d74"} Oct 04 03:15:06 crc kubenswrapper[4742]: I1004 03:15:06.596810 4742 generic.go:334] "Generic (PLEG): container finished" podID="3e50888f-9b21-423e-a450-66f235ad93c8" containerID="1d7fbfb15a25c976db6752e03519fb0040a4eb517aa6a18b24a7b277e0b216fc" exitCode=0 Oct 04 03:15:06 crc kubenswrapper[4742]: I1004 03:15:06.597395 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l5vff" event={"ID":"3e50888f-9b21-423e-a450-66f235ad93c8","Type":"ContainerDied","Data":"1d7fbfb15a25c976db6752e03519fb0040a4eb517aa6a18b24a7b277e0b216fc"} Oct 04 03:15:06 crc kubenswrapper[4742]: I1004 03:15:06.597438 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l5vff" event={"ID":"3e50888f-9b21-423e-a450-66f235ad93c8","Type":"ContainerStarted","Data":"ccc108c20beae3b37a723d22b4c8f15bf0a9a017b180e15fc41a035ac66a424a"} Oct 04 03:15:07 crc kubenswrapper[4742]: I1004 03:15:07.606329 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9ztnx" event={"ID":"f4f312f8-6a5a-41da-9de0-0e67ea5b3344","Type":"ContainerStarted","Data":"175f9abc269fc595de62218a8317c540d85fd5f1a2166c1521b8d123115ccc06"} Oct 04 03:15:07 crc kubenswrapper[4742]: I1004 03:15:07.610284 4742 generic.go:334] "Generic (PLEG): container finished" podID="3e50888f-9b21-423e-a450-66f235ad93c8" containerID="01ba1469fab1493c503c9451af8ae546bb251f0b78bdb7058b3931d32e149e06" exitCode=0 Oct 04 03:15:07 crc kubenswrapper[4742]: I1004 03:15:07.610326 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l5vff" event={"ID":"3e50888f-9b21-423e-a450-66f235ad93c8","Type":"ContainerDied","Data":"01ba1469fab1493c503c9451af8ae546bb251f0b78bdb7058b3931d32e149e06"} Oct 04 03:15:07 crc kubenswrapper[4742]: I1004 03:15:07.627050 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9ztnx" podStartSLOduration=2.22360223 podStartE2EDuration="4.627024498s" podCreationTimestamp="2025-10-04 03:15:03 +0000 UTC" firstStartedPulling="2025-10-04 03:15:04.573174481 +0000 UTC m=+270.505511657" lastFinishedPulling="2025-10-04 03:15:06.976596749 +0000 UTC m=+272.908933925" observedRunningTime="2025-10-04 03:15:07.625846017 +0000 UTC m=+273.558183193" watchObservedRunningTime="2025-10-04 03:15:07.627024498 +0000 UTC m=+273.559361674" Oct 04 03:15:09 crc kubenswrapper[4742]: I1004 03:15:09.622778 4742 generic.go:334] "Generic (PLEG): container finished" podID="4bde2e58-49b8-4975-aaaf-2cecd14c34d1" containerID="078c5f27a0cf95154349db873bcedb76dab9b6889f6bea04bb2573b8ee1f6577" exitCode=0 Oct 04 03:15:09 crc kubenswrapper[4742]: I1004 03:15:09.623207 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t2zpx" event={"ID":"4bde2e58-49b8-4975-aaaf-2cecd14c34d1","Type":"ContainerDied","Data":"078c5f27a0cf95154349db873bcedb76dab9b6889f6bea04bb2573b8ee1f6577"} Oct 04 03:15:09 crc kubenswrapper[4742]: I1004 03:15:09.626891 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l5vff" event={"ID":"3e50888f-9b21-423e-a450-66f235ad93c8","Type":"ContainerStarted","Data":"82db47be8fa07698ec3c776f8c0097fd146f1d9d606e24e8427597c364a63ba3"} Oct 04 03:15:10 crc kubenswrapper[4742]: I1004 03:15:10.632970 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t2zpx" event={"ID":"4bde2e58-49b8-4975-aaaf-2cecd14c34d1","Type":"ContainerStarted","Data":"66eb7f4ca40bc2ad573f8a689cbee21d2c2cc57d70fb99250909c351469cb70c"} Oct 04 03:15:10 crc kubenswrapper[4742]: I1004 03:15:10.651046 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-l5vff" podStartSLOduration=4.224418019 podStartE2EDuration="5.651022755s" podCreationTimestamp="2025-10-04 03:15:05 +0000 UTC" firstStartedPulling="2025-10-04 03:15:06.598580558 +0000 UTC m=+272.530917734" lastFinishedPulling="2025-10-04 03:15:08.025185294 +0000 UTC m=+273.957522470" observedRunningTime="2025-10-04 03:15:09.663417042 +0000 UTC m=+275.595754218" watchObservedRunningTime="2025-10-04 03:15:10.651022755 +0000 UTC m=+276.583359931" Oct 04 03:15:10 crc kubenswrapper[4742]: I1004 03:15:10.652824 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-t2zpx" podStartSLOduration=2.190184441 podStartE2EDuration="5.652813782s" podCreationTimestamp="2025-10-04 03:15:05 +0000 UTC" firstStartedPulling="2025-10-04 03:15:06.590147458 +0000 UTC m=+272.522484634" lastFinishedPulling="2025-10-04 03:15:10.052776809 +0000 UTC m=+275.985113975" observedRunningTime="2025-10-04 03:15:10.65044844 +0000 UTC m=+276.582785616" watchObservedRunningTime="2025-10-04 03:15:10.652813782 +0000 UTC m=+276.585150978" Oct 04 03:15:13 crc kubenswrapper[4742]: I1004 03:15:13.149988 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-c29rn" Oct 04 03:15:13 crc kubenswrapper[4742]: I1004 03:15:13.150355 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-c29rn" Oct 04 03:15:13 crc kubenswrapper[4742]: I1004 03:15:13.187519 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-c29rn" Oct 04 03:15:13 crc kubenswrapper[4742]: I1004 03:15:13.358498 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9ztnx" Oct 04 03:15:13 crc kubenswrapper[4742]: I1004 03:15:13.358585 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9ztnx" Oct 04 03:15:13 crc kubenswrapper[4742]: I1004 03:15:13.394913 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9ztnx" Oct 04 03:15:13 crc kubenswrapper[4742]: I1004 03:15:13.680837 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9ztnx" Oct 04 03:15:13 crc kubenswrapper[4742]: I1004 03:15:13.682582 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-c29rn" Oct 04 03:15:15 crc kubenswrapper[4742]: I1004 03:15:15.632194 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-l5vff" Oct 04 03:15:15 crc kubenswrapper[4742]: I1004 03:15:15.632613 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-l5vff" Oct 04 03:15:15 crc kubenswrapper[4742]: I1004 03:15:15.677860 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-l5vff" Oct 04 03:15:15 crc kubenswrapper[4742]: I1004 03:15:15.716251 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-l5vff" Oct 04 03:15:15 crc kubenswrapper[4742]: I1004 03:15:15.749210 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-t2zpx" Oct 04 03:15:15 crc kubenswrapper[4742]: I1004 03:15:15.749255 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-t2zpx" Oct 04 03:15:15 crc kubenswrapper[4742]: I1004 03:15:15.788083 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-t2zpx" Oct 04 03:15:16 crc kubenswrapper[4742]: I1004 03:15:16.713549 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-t2zpx" Oct 04 03:16:44 crc kubenswrapper[4742]: I1004 03:16:44.872086 4742 patch_prober.go:28] interesting pod/machine-config-daemon-qdt7v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:16:44 crc kubenswrapper[4742]: I1004 03:16:44.872570 4742 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:17:14 crc kubenswrapper[4742]: I1004 03:17:14.872270 4742 patch_prober.go:28] interesting pod/machine-config-daemon-qdt7v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:17:14 crc kubenswrapper[4742]: I1004 03:17:14.872932 4742 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:17:19 crc kubenswrapper[4742]: I1004 03:17:19.658897 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-fpqrm"] Oct 04 03:17:19 crc kubenswrapper[4742]: I1004 03:17:19.660224 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-fpqrm" Oct 04 03:17:19 crc kubenswrapper[4742]: I1004 03:17:19.687618 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-fpqrm"] Oct 04 03:17:19 crc kubenswrapper[4742]: I1004 03:17:19.847574 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-fpqrm\" (UID: \"6422c8f0-922e-4f7c-bd9e-b389ab243f90\") " pod="openshift-image-registry/image-registry-66df7c8f76-fpqrm" Oct 04 03:17:19 crc kubenswrapper[4742]: I1004 03:17:19.847638 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6422c8f0-922e-4f7c-bd9e-b389ab243f90-registry-certificates\") pod \"image-registry-66df7c8f76-fpqrm\" (UID: \"6422c8f0-922e-4f7c-bd9e-b389ab243f90\") " pod="openshift-image-registry/image-registry-66df7c8f76-fpqrm" Oct 04 03:17:19 crc kubenswrapper[4742]: I1004 03:17:19.847663 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6422c8f0-922e-4f7c-bd9e-b389ab243f90-registry-tls\") pod \"image-registry-66df7c8f76-fpqrm\" (UID: \"6422c8f0-922e-4f7c-bd9e-b389ab243f90\") " pod="openshift-image-registry/image-registry-66df7c8f76-fpqrm" Oct 04 03:17:19 crc kubenswrapper[4742]: I1004 03:17:19.847701 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6422c8f0-922e-4f7c-bd9e-b389ab243f90-trusted-ca\") pod \"image-registry-66df7c8f76-fpqrm\" (UID: \"6422c8f0-922e-4f7c-bd9e-b389ab243f90\") " pod="openshift-image-registry/image-registry-66df7c8f76-fpqrm" Oct 04 03:17:19 crc kubenswrapper[4742]: I1004 03:17:19.847731 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6422c8f0-922e-4f7c-bd9e-b389ab243f90-ca-trust-extracted\") pod \"image-registry-66df7c8f76-fpqrm\" (UID: \"6422c8f0-922e-4f7c-bd9e-b389ab243f90\") " pod="openshift-image-registry/image-registry-66df7c8f76-fpqrm" Oct 04 03:17:19 crc kubenswrapper[4742]: I1004 03:17:19.847754 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6422c8f0-922e-4f7c-bd9e-b389ab243f90-bound-sa-token\") pod \"image-registry-66df7c8f76-fpqrm\" (UID: \"6422c8f0-922e-4f7c-bd9e-b389ab243f90\") " pod="openshift-image-registry/image-registry-66df7c8f76-fpqrm" Oct 04 03:17:19 crc kubenswrapper[4742]: I1004 03:17:19.847872 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6422c8f0-922e-4f7c-bd9e-b389ab243f90-installation-pull-secrets\") pod \"image-registry-66df7c8f76-fpqrm\" (UID: \"6422c8f0-922e-4f7c-bd9e-b389ab243f90\") " pod="openshift-image-registry/image-registry-66df7c8f76-fpqrm" Oct 04 03:17:19 crc kubenswrapper[4742]: I1004 03:17:19.847901 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtk5n\" (UniqueName: \"kubernetes.io/projected/6422c8f0-922e-4f7c-bd9e-b389ab243f90-kube-api-access-mtk5n\") pod \"image-registry-66df7c8f76-fpqrm\" (UID: \"6422c8f0-922e-4f7c-bd9e-b389ab243f90\") " pod="openshift-image-registry/image-registry-66df7c8f76-fpqrm" Oct 04 03:17:19 crc kubenswrapper[4742]: I1004 03:17:19.878545 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-fpqrm\" (UID: \"6422c8f0-922e-4f7c-bd9e-b389ab243f90\") " pod="openshift-image-registry/image-registry-66df7c8f76-fpqrm" Oct 04 03:17:19 crc kubenswrapper[4742]: I1004 03:17:19.949577 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6422c8f0-922e-4f7c-bd9e-b389ab243f90-installation-pull-secrets\") pod \"image-registry-66df7c8f76-fpqrm\" (UID: \"6422c8f0-922e-4f7c-bd9e-b389ab243f90\") " pod="openshift-image-registry/image-registry-66df7c8f76-fpqrm" Oct 04 03:17:19 crc kubenswrapper[4742]: I1004 03:17:19.949640 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtk5n\" (UniqueName: \"kubernetes.io/projected/6422c8f0-922e-4f7c-bd9e-b389ab243f90-kube-api-access-mtk5n\") pod \"image-registry-66df7c8f76-fpqrm\" (UID: \"6422c8f0-922e-4f7c-bd9e-b389ab243f90\") " pod="openshift-image-registry/image-registry-66df7c8f76-fpqrm" Oct 04 03:17:19 crc kubenswrapper[4742]: I1004 03:17:19.949679 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6422c8f0-922e-4f7c-bd9e-b389ab243f90-registry-certificates\") pod \"image-registry-66df7c8f76-fpqrm\" (UID: \"6422c8f0-922e-4f7c-bd9e-b389ab243f90\") " pod="openshift-image-registry/image-registry-66df7c8f76-fpqrm" Oct 04 03:17:19 crc kubenswrapper[4742]: I1004 03:17:19.949703 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6422c8f0-922e-4f7c-bd9e-b389ab243f90-registry-tls\") pod \"image-registry-66df7c8f76-fpqrm\" (UID: \"6422c8f0-922e-4f7c-bd9e-b389ab243f90\") " pod="openshift-image-registry/image-registry-66df7c8f76-fpqrm" Oct 04 03:17:19 crc kubenswrapper[4742]: I1004 03:17:19.949749 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6422c8f0-922e-4f7c-bd9e-b389ab243f90-trusted-ca\") pod \"image-registry-66df7c8f76-fpqrm\" (UID: \"6422c8f0-922e-4f7c-bd9e-b389ab243f90\") " pod="openshift-image-registry/image-registry-66df7c8f76-fpqrm" Oct 04 03:17:19 crc kubenswrapper[4742]: I1004 03:17:19.949777 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6422c8f0-922e-4f7c-bd9e-b389ab243f90-ca-trust-extracted\") pod \"image-registry-66df7c8f76-fpqrm\" (UID: \"6422c8f0-922e-4f7c-bd9e-b389ab243f90\") " pod="openshift-image-registry/image-registry-66df7c8f76-fpqrm" Oct 04 03:17:19 crc kubenswrapper[4742]: I1004 03:17:19.949804 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6422c8f0-922e-4f7c-bd9e-b389ab243f90-bound-sa-token\") pod \"image-registry-66df7c8f76-fpqrm\" (UID: \"6422c8f0-922e-4f7c-bd9e-b389ab243f90\") " pod="openshift-image-registry/image-registry-66df7c8f76-fpqrm" Oct 04 03:17:19 crc kubenswrapper[4742]: I1004 03:17:19.951735 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6422c8f0-922e-4f7c-bd9e-b389ab243f90-trusted-ca\") pod \"image-registry-66df7c8f76-fpqrm\" (UID: \"6422c8f0-922e-4f7c-bd9e-b389ab243f90\") " pod="openshift-image-registry/image-registry-66df7c8f76-fpqrm" Oct 04 03:17:19 crc kubenswrapper[4742]: I1004 03:17:19.952105 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6422c8f0-922e-4f7c-bd9e-b389ab243f90-ca-trust-extracted\") pod \"image-registry-66df7c8f76-fpqrm\" (UID: \"6422c8f0-922e-4f7c-bd9e-b389ab243f90\") " pod="openshift-image-registry/image-registry-66df7c8f76-fpqrm" Oct 04 03:17:19 crc kubenswrapper[4742]: I1004 03:17:19.953787 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6422c8f0-922e-4f7c-bd9e-b389ab243f90-registry-certificates\") pod \"image-registry-66df7c8f76-fpqrm\" (UID: \"6422c8f0-922e-4f7c-bd9e-b389ab243f90\") " pod="openshift-image-registry/image-registry-66df7c8f76-fpqrm" Oct 04 03:17:19 crc kubenswrapper[4742]: I1004 03:17:19.958833 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6422c8f0-922e-4f7c-bd9e-b389ab243f90-registry-tls\") pod \"image-registry-66df7c8f76-fpqrm\" (UID: \"6422c8f0-922e-4f7c-bd9e-b389ab243f90\") " pod="openshift-image-registry/image-registry-66df7c8f76-fpqrm" Oct 04 03:17:19 crc kubenswrapper[4742]: I1004 03:17:19.961107 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6422c8f0-922e-4f7c-bd9e-b389ab243f90-installation-pull-secrets\") pod \"image-registry-66df7c8f76-fpqrm\" (UID: \"6422c8f0-922e-4f7c-bd9e-b389ab243f90\") " pod="openshift-image-registry/image-registry-66df7c8f76-fpqrm" Oct 04 03:17:19 crc kubenswrapper[4742]: I1004 03:17:19.967112 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtk5n\" (UniqueName: \"kubernetes.io/projected/6422c8f0-922e-4f7c-bd9e-b389ab243f90-kube-api-access-mtk5n\") pod \"image-registry-66df7c8f76-fpqrm\" (UID: \"6422c8f0-922e-4f7c-bd9e-b389ab243f90\") " pod="openshift-image-registry/image-registry-66df7c8f76-fpqrm" Oct 04 03:17:19 crc kubenswrapper[4742]: I1004 03:17:19.971651 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6422c8f0-922e-4f7c-bd9e-b389ab243f90-bound-sa-token\") pod \"image-registry-66df7c8f76-fpqrm\" (UID: \"6422c8f0-922e-4f7c-bd9e-b389ab243f90\") " pod="openshift-image-registry/image-registry-66df7c8f76-fpqrm" Oct 04 03:17:19 crc kubenswrapper[4742]: I1004 03:17:19.990515 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-fpqrm" Oct 04 03:17:20 crc kubenswrapper[4742]: I1004 03:17:20.247572 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-fpqrm"] Oct 04 03:17:20 crc kubenswrapper[4742]: I1004 03:17:20.408944 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-fpqrm" event={"ID":"6422c8f0-922e-4f7c-bd9e-b389ab243f90","Type":"ContainerStarted","Data":"3d702dd50ae9e87b35ac6ffe8482ba0a92bda18c3318db0086a55b21ca44604f"} Oct 04 03:17:20 crc kubenswrapper[4742]: I1004 03:17:20.409986 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-fpqrm" event={"ID":"6422c8f0-922e-4f7c-bd9e-b389ab243f90","Type":"ContainerStarted","Data":"468068b5245abb032fbbc1b42c1385556994939fcd2de0c90153e6aeb161eb55"} Oct 04 03:17:20 crc kubenswrapper[4742]: I1004 03:17:20.410413 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-fpqrm" Oct 04 03:17:20 crc kubenswrapper[4742]: I1004 03:17:20.434306 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-fpqrm" podStartSLOduration=1.434256375 podStartE2EDuration="1.434256375s" podCreationTimestamp="2025-10-04 03:17:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:17:20.43333901 +0000 UTC m=+406.365676196" watchObservedRunningTime="2025-10-04 03:17:20.434256375 +0000 UTC m=+406.366593561" Oct 04 03:17:39 crc kubenswrapper[4742]: I1004 03:17:39.996475 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-fpqrm" Oct 04 03:17:40 crc kubenswrapper[4742]: I1004 03:17:40.048022 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-f6gcs"] Oct 04 03:17:44 crc kubenswrapper[4742]: I1004 03:17:44.871761 4742 patch_prober.go:28] interesting pod/machine-config-daemon-qdt7v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:17:44 crc kubenswrapper[4742]: I1004 03:17:44.872361 4742 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:17:44 crc kubenswrapper[4742]: I1004 03:17:44.872470 4742 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" Oct 04 03:17:44 crc kubenswrapper[4742]: I1004 03:17:44.873383 4742 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"265d94a26e57f0eae37e63999ff3e9ea88f239556449abe4f5a9ca7688a659d5"} pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 03:17:44 crc kubenswrapper[4742]: I1004 03:17:44.873495 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerName="machine-config-daemon" containerID="cri-o://265d94a26e57f0eae37e63999ff3e9ea88f239556449abe4f5a9ca7688a659d5" gracePeriod=600 Oct 04 03:17:45 crc kubenswrapper[4742]: I1004 03:17:45.574210 4742 generic.go:334] "Generic (PLEG): container finished" podID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerID="265d94a26e57f0eae37e63999ff3e9ea88f239556449abe4f5a9ca7688a659d5" exitCode=0 Oct 04 03:17:45 crc kubenswrapper[4742]: I1004 03:17:45.574293 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" event={"ID":"989ea247-7ac9-4ade-8dcb-2632fef4f83e","Type":"ContainerDied","Data":"265d94a26e57f0eae37e63999ff3e9ea88f239556449abe4f5a9ca7688a659d5"} Oct 04 03:17:45 crc kubenswrapper[4742]: I1004 03:17:45.574815 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" event={"ID":"989ea247-7ac9-4ade-8dcb-2632fef4f83e","Type":"ContainerStarted","Data":"716e86aaee18fb777354d280f34d7ae9aea9909aae176a0c4c726413f1e1cb06"} Oct 04 03:17:45 crc kubenswrapper[4742]: I1004 03:17:45.574854 4742 scope.go:117] "RemoveContainer" containerID="dd206f68e352a5a3bcef60387b30ac1848e5c8714697805a5b9728ba8cc4fbde" Oct 04 03:18:05 crc kubenswrapper[4742]: I1004 03:18:05.095586 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" podUID="4892c4de-41a2-4c58-bb59-36a6c584da2f" containerName="registry" containerID="cri-o://8da953400f2531c9e7ce60b4d31e657c743b0c1b885feb997298d47edaf94623" gracePeriod=30 Oct 04 03:18:05 crc kubenswrapper[4742]: I1004 03:18:05.476379 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:18:05 crc kubenswrapper[4742]: I1004 03:18:05.670616 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4892c4de-41a2-4c58-bb59-36a6c584da2f-ca-trust-extracted\") pod \"4892c4de-41a2-4c58-bb59-36a6c584da2f\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " Oct 04 03:18:05 crc kubenswrapper[4742]: I1004 03:18:05.671067 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4892c4de-41a2-4c58-bb59-36a6c584da2f-registry-certificates\") pod \"4892c4de-41a2-4c58-bb59-36a6c584da2f\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " Oct 04 03:18:05 crc kubenswrapper[4742]: I1004 03:18:05.671110 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk46r\" (UniqueName: \"kubernetes.io/projected/4892c4de-41a2-4c58-bb59-36a6c584da2f-kube-api-access-tk46r\") pod \"4892c4de-41a2-4c58-bb59-36a6c584da2f\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " Oct 04 03:18:05 crc kubenswrapper[4742]: I1004 03:18:05.671359 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"4892c4de-41a2-4c58-bb59-36a6c584da2f\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " Oct 04 03:18:05 crc kubenswrapper[4742]: I1004 03:18:05.671381 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4892c4de-41a2-4c58-bb59-36a6c584da2f-installation-pull-secrets\") pod \"4892c4de-41a2-4c58-bb59-36a6c584da2f\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " Oct 04 03:18:05 crc kubenswrapper[4742]: I1004 03:18:05.671403 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4892c4de-41a2-4c58-bb59-36a6c584da2f-bound-sa-token\") pod \"4892c4de-41a2-4c58-bb59-36a6c584da2f\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " Oct 04 03:18:05 crc kubenswrapper[4742]: I1004 03:18:05.671422 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4892c4de-41a2-4c58-bb59-36a6c584da2f-trusted-ca\") pod \"4892c4de-41a2-4c58-bb59-36a6c584da2f\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " Oct 04 03:18:05 crc kubenswrapper[4742]: I1004 03:18:05.671499 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4892c4de-41a2-4c58-bb59-36a6c584da2f-registry-tls\") pod \"4892c4de-41a2-4c58-bb59-36a6c584da2f\" (UID: \"4892c4de-41a2-4c58-bb59-36a6c584da2f\") " Oct 04 03:18:05 crc kubenswrapper[4742]: I1004 03:18:05.673243 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4892c4de-41a2-4c58-bb59-36a6c584da2f-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "4892c4de-41a2-4c58-bb59-36a6c584da2f" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:18:05 crc kubenswrapper[4742]: I1004 03:18:05.675067 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4892c4de-41a2-4c58-bb59-36a6c584da2f-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "4892c4de-41a2-4c58-bb59-36a6c584da2f" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:18:05 crc kubenswrapper[4742]: I1004 03:18:05.679895 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4892c4de-41a2-4c58-bb59-36a6c584da2f-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "4892c4de-41a2-4c58-bb59-36a6c584da2f" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:18:05 crc kubenswrapper[4742]: I1004 03:18:05.680625 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4892c4de-41a2-4c58-bb59-36a6c584da2f-kube-api-access-tk46r" (OuterVolumeSpecName: "kube-api-access-tk46r") pod "4892c4de-41a2-4c58-bb59-36a6c584da2f" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f"). InnerVolumeSpecName "kube-api-access-tk46r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:18:05 crc kubenswrapper[4742]: I1004 03:18:05.680846 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4892c4de-41a2-4c58-bb59-36a6c584da2f-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "4892c4de-41a2-4c58-bb59-36a6c584da2f" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:18:05 crc kubenswrapper[4742]: I1004 03:18:05.681720 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4892c4de-41a2-4c58-bb59-36a6c584da2f-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "4892c4de-41a2-4c58-bb59-36a6c584da2f" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:18:05 crc kubenswrapper[4742]: I1004 03:18:05.691440 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "4892c4de-41a2-4c58-bb59-36a6c584da2f" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 04 03:18:05 crc kubenswrapper[4742]: I1004 03:18:05.694207 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4892c4de-41a2-4c58-bb59-36a6c584da2f-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "4892c4de-41a2-4c58-bb59-36a6c584da2f" (UID: "4892c4de-41a2-4c58-bb59-36a6c584da2f"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:18:05 crc kubenswrapper[4742]: I1004 03:18:05.731674 4742 generic.go:334] "Generic (PLEG): container finished" podID="4892c4de-41a2-4c58-bb59-36a6c584da2f" containerID="8da953400f2531c9e7ce60b4d31e657c743b0c1b885feb997298d47edaf94623" exitCode=0 Oct 04 03:18:05 crc kubenswrapper[4742]: I1004 03:18:05.731727 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" event={"ID":"4892c4de-41a2-4c58-bb59-36a6c584da2f","Type":"ContainerDied","Data":"8da953400f2531c9e7ce60b4d31e657c743b0c1b885feb997298d47edaf94623"} Oct 04 03:18:05 crc kubenswrapper[4742]: I1004 03:18:05.731767 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" event={"ID":"4892c4de-41a2-4c58-bb59-36a6c584da2f","Type":"ContainerDied","Data":"c410b0070a34c78d7e797e2b37e127fc41594f11f25c72577395a93be0a6cc03"} Oct 04 03:18:05 crc kubenswrapper[4742]: I1004 03:18:05.731786 4742 scope.go:117] "RemoveContainer" containerID="8da953400f2531c9e7ce60b4d31e657c743b0c1b885feb997298d47edaf94623" Oct 04 03:18:05 crc kubenswrapper[4742]: I1004 03:18:05.731808 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-f6gcs" Oct 04 03:18:05 crc kubenswrapper[4742]: I1004 03:18:05.766023 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-f6gcs"] Oct 04 03:18:05 crc kubenswrapper[4742]: I1004 03:18:05.766614 4742 scope.go:117] "RemoveContainer" containerID="8da953400f2531c9e7ce60b4d31e657c743b0c1b885feb997298d47edaf94623" Oct 04 03:18:05 crc kubenswrapper[4742]: E1004 03:18:05.767506 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8da953400f2531c9e7ce60b4d31e657c743b0c1b885feb997298d47edaf94623\": container with ID starting with 8da953400f2531c9e7ce60b4d31e657c743b0c1b885feb997298d47edaf94623 not found: ID does not exist" containerID="8da953400f2531c9e7ce60b4d31e657c743b0c1b885feb997298d47edaf94623" Oct 04 03:18:05 crc kubenswrapper[4742]: I1004 03:18:05.767537 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8da953400f2531c9e7ce60b4d31e657c743b0c1b885feb997298d47edaf94623"} err="failed to get container status \"8da953400f2531c9e7ce60b4d31e657c743b0c1b885feb997298d47edaf94623\": rpc error: code = NotFound desc = could not find container \"8da953400f2531c9e7ce60b4d31e657c743b0c1b885feb997298d47edaf94623\": container with ID starting with 8da953400f2531c9e7ce60b4d31e657c743b0c1b885feb997298d47edaf94623 not found: ID does not exist" Oct 04 03:18:05 crc kubenswrapper[4742]: I1004 03:18:05.774076 4742 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4892c4de-41a2-4c58-bb59-36a6c584da2f-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 04 03:18:05 crc kubenswrapper[4742]: I1004 03:18:05.774141 4742 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4892c4de-41a2-4c58-bb59-36a6c584da2f-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 04 03:18:05 crc kubenswrapper[4742]: I1004 03:18:05.774165 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk46r\" (UniqueName: \"kubernetes.io/projected/4892c4de-41a2-4c58-bb59-36a6c584da2f-kube-api-access-tk46r\") on node \"crc\" DevicePath \"\"" Oct 04 03:18:05 crc kubenswrapper[4742]: I1004 03:18:05.774190 4742 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4892c4de-41a2-4c58-bb59-36a6c584da2f-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 04 03:18:05 crc kubenswrapper[4742]: I1004 03:18:05.774210 4742 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4892c4de-41a2-4c58-bb59-36a6c584da2f-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 04 03:18:05 crc kubenswrapper[4742]: I1004 03:18:05.774230 4742 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4892c4de-41a2-4c58-bb59-36a6c584da2f-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:18:05 crc kubenswrapper[4742]: I1004 03:18:05.774248 4742 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4892c4de-41a2-4c58-bb59-36a6c584da2f-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:18:05 crc kubenswrapper[4742]: I1004 03:18:05.774542 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-f6gcs"] Oct 04 03:18:06 crc kubenswrapper[4742]: I1004 03:18:06.870950 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4892c4de-41a2-4c58-bb59-36a6c584da2f" path="/var/lib/kubelet/pods/4892c4de-41a2-4c58-bb59-36a6c584da2f/volumes" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.298853 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9qrtl"] Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.300225 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="ovn-controller" containerID="cri-o://78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24" gracePeriod=30 Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.300847 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="sbdb" containerID="cri-o://e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4" gracePeriod=30 Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.300915 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="nbdb" containerID="cri-o://0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433" gracePeriod=30 Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.300966 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="northd" containerID="cri-o://e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6" gracePeriod=30 Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.301029 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236" gracePeriod=30 Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.301085 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="kube-rbac-proxy-node" containerID="cri-o://ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993" gracePeriod=30 Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.301141 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="ovn-acl-logging" containerID="cri-o://73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b" gracePeriod=30 Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.357531 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="ovnkube-controller" containerID="cri-o://8bf4cc53b1c45b02faf41e5d40191f4fc7ff1959030bfa049955d8e7a31350e7" gracePeriod=30 Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.509214 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nc8m7_46de52ee-226c-4713-bd7b-961f8f43f635/kube-multus/2.log" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.509692 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nc8m7_46de52ee-226c-4713-bd7b-961f8f43f635/kube-multus/1.log" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.509722 4742 generic.go:334] "Generic (PLEG): container finished" podID="46de52ee-226c-4713-bd7b-961f8f43f635" containerID="7a1d02fda520e7052329a1d1c9e5516ad70f170819a7d5263271e87b5928ba06" exitCode=2 Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.509775 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nc8m7" event={"ID":"46de52ee-226c-4713-bd7b-961f8f43f635","Type":"ContainerDied","Data":"7a1d02fda520e7052329a1d1c9e5516ad70f170819a7d5263271e87b5928ba06"} Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.509818 4742 scope.go:117] "RemoveContainer" containerID="bfe5cd4a2153e5ae33beedc1cfdee7c39f57d2dc41a3d5e7d442df75d3c069db" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.510266 4742 scope.go:117] "RemoveContainer" containerID="7a1d02fda520e7052329a1d1c9e5516ad70f170819a7d5263271e87b5928ba06" Oct 04 03:20:14 crc kubenswrapper[4742]: E1004 03:20:14.510547 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-nc8m7_openshift-multus(46de52ee-226c-4713-bd7b-961f8f43f635)\"" pod="openshift-multus/multus-nc8m7" podUID="46de52ee-226c-4713-bd7b-961f8f43f635" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.514015 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qrtl_0cfc6e64-ba08-4c44-a18e-3b93ab792a6a/ovnkube-controller/3.log" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.518568 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qrtl_0cfc6e64-ba08-4c44-a18e-3b93ab792a6a/ovn-acl-logging/0.log" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.519073 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qrtl_0cfc6e64-ba08-4c44-a18e-3b93ab792a6a/ovn-controller/0.log" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.519603 4742 generic.go:334] "Generic (PLEG): container finished" podID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerID="8bf4cc53b1c45b02faf41e5d40191f4fc7ff1959030bfa049955d8e7a31350e7" exitCode=0 Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.519638 4742 generic.go:334] "Generic (PLEG): container finished" podID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerID="ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236" exitCode=0 Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.519647 4742 generic.go:334] "Generic (PLEG): container finished" podID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerID="ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993" exitCode=0 Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.519655 4742 generic.go:334] "Generic (PLEG): container finished" podID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerID="73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b" exitCode=143 Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.519662 4742 generic.go:334] "Generic (PLEG): container finished" podID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerID="78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24" exitCode=143 Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.519700 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" event={"ID":"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a","Type":"ContainerDied","Data":"8bf4cc53b1c45b02faf41e5d40191f4fc7ff1959030bfa049955d8e7a31350e7"} Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.519749 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" event={"ID":"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a","Type":"ContainerDied","Data":"ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236"} Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.519766 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" event={"ID":"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a","Type":"ContainerDied","Data":"ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993"} Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.519780 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" event={"ID":"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a","Type":"ContainerDied","Data":"73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b"} Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.519791 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" event={"ID":"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a","Type":"ContainerDied","Data":"78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24"} Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.636129 4742 scope.go:117] "RemoveContainer" containerID="9b37c3bfb97a4dc710d7664dfa0bc1cd0093419f21a8ca483ac1f38e695d7271" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.644826 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qrtl_0cfc6e64-ba08-4c44-a18e-3b93ab792a6a/ovn-acl-logging/0.log" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.645600 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qrtl_0cfc6e64-ba08-4c44-a18e-3b93ab792a6a/ovn-controller/0.log" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.646115 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.680676 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-cni-bin\") pod \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.680725 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-var-lib-openvswitch\") pod \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.680753 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-systemd-units\") pod \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.680781 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-ovnkube-config\") pod \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.680801 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-run-ovn-kubernetes\") pod \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.680823 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-run-openvswitch\") pod \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.680841 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-cni-netd\") pod \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.680859 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-env-overrides\") pod \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.680882 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.680909 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-run-ovn\") pod \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.680930 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-etc-openvswitch\") pod \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.680952 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-log-socket\") pod \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.680969 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-node-log\") pod \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.680986 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-kubelet\") pod \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.681010 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-run-systemd\") pod \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.681031 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-slash\") pod \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.681054 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-run-netns\") pod \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.681072 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txjzp\" (UniqueName: \"kubernetes.io/projected/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-kube-api-access-txjzp\") pod \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.681090 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-ovnkube-script-lib\") pod \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.681110 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-ovn-node-metrics-cert\") pod \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\" (UID: \"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a\") " Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.680946 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" (UID: "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.681014 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" (UID: "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.681019 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" (UID: "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.681042 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" (UID: "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.681052 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-node-log" (OuterVolumeSpecName: "node-log") pod "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" (UID: "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.681072 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" (UID: "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.681086 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" (UID: "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.681099 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-log-socket" (OuterVolumeSpecName: "log-socket") pod "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" (UID: "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.681121 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" (UID: "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.681135 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" (UID: "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.681154 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-slash" (OuterVolumeSpecName: "host-slash") pod "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" (UID: "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.681166 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" (UID: "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.681299 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" (UID: "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.681662 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" (UID: "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.681702 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" (UID: "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.682054 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" (UID: "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.682668 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" (UID: "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.683034 4742 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.683063 4742 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.683076 4742 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.683086 4742 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.683098 4742 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.683109 4742 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.683117 4742 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.683130 4742 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.683141 4742 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.683152 4742 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.683162 4742 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.683175 4742 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-log-socket\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.683184 4742 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-node-log\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.683193 4742 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.683203 4742 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-slash\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.683214 4742 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.683224 4742 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.687762 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" (UID: "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.690308 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-kube-api-access-txjzp" (OuterVolumeSpecName: "kube-api-access-txjzp") pod "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" (UID: "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a"). InnerVolumeSpecName "kube-api-access-txjzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.694369 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" (UID: "0cfc6e64-ba08-4c44-a18e-3b93ab792a6a"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.703207 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-ql4ht"] Oct 04 03:20:14 crc kubenswrapper[4742]: E1004 03:20:14.703413 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="ovn-controller" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.703426 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="ovn-controller" Oct 04 03:20:14 crc kubenswrapper[4742]: E1004 03:20:14.703437 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="ovn-acl-logging" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.703444 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="ovn-acl-logging" Oct 04 03:20:14 crc kubenswrapper[4742]: E1004 03:20:14.703451 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="kube-rbac-proxy-ovn-metrics" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.703459 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="kube-rbac-proxy-ovn-metrics" Oct 04 03:20:14 crc kubenswrapper[4742]: E1004 03:20:14.703465 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="sbdb" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.703471 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="sbdb" Oct 04 03:20:14 crc kubenswrapper[4742]: E1004 03:20:14.703480 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="ovnkube-controller" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.703487 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="ovnkube-controller" Oct 04 03:20:14 crc kubenswrapper[4742]: E1004 03:20:14.703495 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="ovnkube-controller" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.703500 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="ovnkube-controller" Oct 04 03:20:14 crc kubenswrapper[4742]: E1004 03:20:14.703508 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="kube-rbac-proxy-node" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.703514 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="kube-rbac-proxy-node" Oct 04 03:20:14 crc kubenswrapper[4742]: E1004 03:20:14.703524 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="ovnkube-controller" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.703531 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="ovnkube-controller" Oct 04 03:20:14 crc kubenswrapper[4742]: E1004 03:20:14.703539 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="northd" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.703546 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="northd" Oct 04 03:20:14 crc kubenswrapper[4742]: E1004 03:20:14.703555 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="kubecfg-setup" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.703561 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="kubecfg-setup" Oct 04 03:20:14 crc kubenswrapper[4742]: E1004 03:20:14.703570 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4892c4de-41a2-4c58-bb59-36a6c584da2f" containerName="registry" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.703576 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="4892c4de-41a2-4c58-bb59-36a6c584da2f" containerName="registry" Oct 04 03:20:14 crc kubenswrapper[4742]: E1004 03:20:14.703585 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="nbdb" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.703591 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="nbdb" Oct 04 03:20:14 crc kubenswrapper[4742]: E1004 03:20:14.703598 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="ovnkube-controller" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.703604 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="ovnkube-controller" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.703682 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="ovnkube-controller" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.703691 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="ovnkube-controller" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.703698 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="northd" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.703706 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="ovn-controller" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.703713 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="ovnkube-controller" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.703722 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="4892c4de-41a2-4c58-bb59-36a6c584da2f" containerName="registry" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.703730 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="kube-rbac-proxy-ovn-metrics" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.703737 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="ovn-acl-logging" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.703745 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="sbdb" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.703751 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="kube-rbac-proxy-node" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.703759 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="nbdb" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.703766 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="ovnkube-controller" Oct 04 03:20:14 crc kubenswrapper[4742]: E1004 03:20:14.703850 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="ovnkube-controller" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.703856 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="ovnkube-controller" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.703935 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerName="ovnkube-controller" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.705819 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.784724 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.784790 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e331ee65-e461-4722-8ed4-8af04bd0743c-ovnkube-script-lib\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.784811 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-etc-openvswitch\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.784829 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-host-cni-netd\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.784849 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-var-lib-openvswitch\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.784868 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-host-kubelet\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.784933 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e331ee65-e461-4722-8ed4-8af04bd0743c-ovnkube-config\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.784954 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e331ee65-e461-4722-8ed4-8af04bd0743c-env-overrides\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.784968 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-log-socket\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.784988 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-host-run-netns\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.785030 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-systemd-units\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.785047 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e331ee65-e461-4722-8ed4-8af04bd0743c-ovn-node-metrics-cert\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.785064 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-run-openvswitch\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.785078 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-host-cni-bin\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.785141 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-host-run-ovn-kubernetes\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.785191 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q2zf\" (UniqueName: \"kubernetes.io/projected/e331ee65-e461-4722-8ed4-8af04bd0743c-kube-api-access-8q2zf\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.785238 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-host-slash\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.785321 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-run-systemd\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.785361 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-run-ovn\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.785378 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-node-log\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.785440 4742 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.785451 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txjzp\" (UniqueName: \"kubernetes.io/projected/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-kube-api-access-txjzp\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.785460 4742 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.872233 4742 patch_prober.go:28] interesting pod/machine-config-daemon-qdt7v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.872562 4742 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.886247 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-run-openvswitch\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.886329 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-host-cni-bin\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.886362 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-host-run-ovn-kubernetes\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.886408 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8q2zf\" (UniqueName: \"kubernetes.io/projected/e331ee65-e461-4722-8ed4-8af04bd0743c-kube-api-access-8q2zf\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.886442 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-host-slash\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.886475 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-run-systemd\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.886516 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-run-ovn\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.886541 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-node-log\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.886575 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.886609 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-etc-openvswitch\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.886630 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-host-cni-netd\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.886654 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e331ee65-e461-4722-8ed4-8af04bd0743c-ovnkube-script-lib\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.886680 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-var-lib-openvswitch\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.886705 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-host-kubelet\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.886741 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e331ee65-e461-4722-8ed4-8af04bd0743c-ovnkube-config\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.886793 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e331ee65-e461-4722-8ed4-8af04bd0743c-env-overrides\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.886818 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-log-socket\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.886860 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-host-run-netns\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.886970 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-etc-openvswitch\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.887012 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-host-slash\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.887029 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-run-systemd\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.887061 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-run-openvswitch\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.887073 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-run-ovn\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.887091 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-host-cni-bin\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.887107 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-node-log\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.887121 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-host-run-ovn-kubernetes\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.887146 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.887181 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-var-lib-openvswitch\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.887210 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-host-kubelet\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.887241 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-log-socket\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.887367 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-systemd-units\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.887406 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-host-cni-netd\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.887415 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e331ee65-e461-4722-8ed4-8af04bd0743c-ovn-node-metrics-cert\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.887440 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-host-run-netns\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.887695 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e331ee65-e461-4722-8ed4-8af04bd0743c-ovnkube-config\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.887732 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e331ee65-e461-4722-8ed4-8af04bd0743c-systemd-units\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.887895 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e331ee65-e461-4722-8ed4-8af04bd0743c-ovnkube-script-lib\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.888804 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e331ee65-e461-4722-8ed4-8af04bd0743c-env-overrides\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.892401 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e331ee65-e461-4722-8ed4-8af04bd0743c-ovn-node-metrics-cert\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:14 crc kubenswrapper[4742]: I1004 03:20:14.905683 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8q2zf\" (UniqueName: \"kubernetes.io/projected/e331ee65-e461-4722-8ed4-8af04bd0743c-kube-api-access-8q2zf\") pod \"ovnkube-node-ql4ht\" (UID: \"e331ee65-e461-4722-8ed4-8af04bd0743c\") " pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.019647 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.530420 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nc8m7_46de52ee-226c-4713-bd7b-961f8f43f635/kube-multus/2.log" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.532570 4742 generic.go:334] "Generic (PLEG): container finished" podID="e331ee65-e461-4722-8ed4-8af04bd0743c" containerID="96ddc3c0d1edcecd987484034794f287c9c1f22728b23e9a101024354f2937ae" exitCode=0 Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.532643 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" event={"ID":"e331ee65-e461-4722-8ed4-8af04bd0743c","Type":"ContainerDied","Data":"96ddc3c0d1edcecd987484034794f287c9c1f22728b23e9a101024354f2937ae"} Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.532721 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" event={"ID":"e331ee65-e461-4722-8ed4-8af04bd0743c","Type":"ContainerStarted","Data":"cc0fd2494094882db7d8d1cf650737b469430d64d1a7cd31297fd1971fe00f08"} Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.540051 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qrtl_0cfc6e64-ba08-4c44-a18e-3b93ab792a6a/ovn-acl-logging/0.log" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.540821 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9qrtl_0cfc6e64-ba08-4c44-a18e-3b93ab792a6a/ovn-controller/0.log" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.541319 4742 generic.go:334] "Generic (PLEG): container finished" podID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerID="e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4" exitCode=0 Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.541357 4742 generic.go:334] "Generic (PLEG): container finished" podID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerID="0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433" exitCode=0 Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.541365 4742 generic.go:334] "Generic (PLEG): container finished" podID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" containerID="e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6" exitCode=0 Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.541391 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" event={"ID":"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a","Type":"ContainerDied","Data":"e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4"} Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.541451 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" event={"ID":"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a","Type":"ContainerDied","Data":"0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433"} Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.541466 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" event={"ID":"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a","Type":"ContainerDied","Data":"e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6"} Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.541477 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" event={"ID":"0cfc6e64-ba08-4c44-a18e-3b93ab792a6a","Type":"ContainerDied","Data":"fbd3b1a032009427bfc30386ecd659463d8b4e86e8278e906adb3c7f353e9e91"} Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.541499 4742 scope.go:117] "RemoveContainer" containerID="8bf4cc53b1c45b02faf41e5d40191f4fc7ff1959030bfa049955d8e7a31350e7" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.541957 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9qrtl" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.557821 4742 scope.go:117] "RemoveContainer" containerID="e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.609553 4742 scope.go:117] "RemoveContainer" containerID="0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.615823 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9qrtl"] Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.616561 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9qrtl"] Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.645859 4742 scope.go:117] "RemoveContainer" containerID="e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.661398 4742 scope.go:117] "RemoveContainer" containerID="ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.675939 4742 scope.go:117] "RemoveContainer" containerID="ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.691381 4742 scope.go:117] "RemoveContainer" containerID="73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.709218 4742 scope.go:117] "RemoveContainer" containerID="78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.727026 4742 scope.go:117] "RemoveContainer" containerID="53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.761999 4742 scope.go:117] "RemoveContainer" containerID="8bf4cc53b1c45b02faf41e5d40191f4fc7ff1959030bfa049955d8e7a31350e7" Oct 04 03:20:15 crc kubenswrapper[4742]: E1004 03:20:15.763140 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8bf4cc53b1c45b02faf41e5d40191f4fc7ff1959030bfa049955d8e7a31350e7\": container with ID starting with 8bf4cc53b1c45b02faf41e5d40191f4fc7ff1959030bfa049955d8e7a31350e7 not found: ID does not exist" containerID="8bf4cc53b1c45b02faf41e5d40191f4fc7ff1959030bfa049955d8e7a31350e7" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.763180 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bf4cc53b1c45b02faf41e5d40191f4fc7ff1959030bfa049955d8e7a31350e7"} err="failed to get container status \"8bf4cc53b1c45b02faf41e5d40191f4fc7ff1959030bfa049955d8e7a31350e7\": rpc error: code = NotFound desc = could not find container \"8bf4cc53b1c45b02faf41e5d40191f4fc7ff1959030bfa049955d8e7a31350e7\": container with ID starting with 8bf4cc53b1c45b02faf41e5d40191f4fc7ff1959030bfa049955d8e7a31350e7 not found: ID does not exist" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.763210 4742 scope.go:117] "RemoveContainer" containerID="e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4" Oct 04 03:20:15 crc kubenswrapper[4742]: E1004 03:20:15.763869 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4\": container with ID starting with e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4 not found: ID does not exist" containerID="e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.763907 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4"} err="failed to get container status \"e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4\": rpc error: code = NotFound desc = could not find container \"e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4\": container with ID starting with e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4 not found: ID does not exist" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.763935 4742 scope.go:117] "RemoveContainer" containerID="0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433" Oct 04 03:20:15 crc kubenswrapper[4742]: E1004 03:20:15.764402 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433\": container with ID starting with 0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433 not found: ID does not exist" containerID="0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.764469 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433"} err="failed to get container status \"0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433\": rpc error: code = NotFound desc = could not find container \"0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433\": container with ID starting with 0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433 not found: ID does not exist" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.764526 4742 scope.go:117] "RemoveContainer" containerID="e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6" Oct 04 03:20:15 crc kubenswrapper[4742]: E1004 03:20:15.764925 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6\": container with ID starting with e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6 not found: ID does not exist" containerID="e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.764963 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6"} err="failed to get container status \"e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6\": rpc error: code = NotFound desc = could not find container \"e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6\": container with ID starting with e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6 not found: ID does not exist" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.764985 4742 scope.go:117] "RemoveContainer" containerID="ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236" Oct 04 03:20:15 crc kubenswrapper[4742]: E1004 03:20:15.765317 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236\": container with ID starting with ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236 not found: ID does not exist" containerID="ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.765345 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236"} err="failed to get container status \"ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236\": rpc error: code = NotFound desc = could not find container \"ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236\": container with ID starting with ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236 not found: ID does not exist" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.765363 4742 scope.go:117] "RemoveContainer" containerID="ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993" Oct 04 03:20:15 crc kubenswrapper[4742]: E1004 03:20:15.765701 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993\": container with ID starting with ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993 not found: ID does not exist" containerID="ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.765775 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993"} err="failed to get container status \"ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993\": rpc error: code = NotFound desc = could not find container \"ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993\": container with ID starting with ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993 not found: ID does not exist" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.765825 4742 scope.go:117] "RemoveContainer" containerID="73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b" Oct 04 03:20:15 crc kubenswrapper[4742]: E1004 03:20:15.766305 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b\": container with ID starting with 73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b not found: ID does not exist" containerID="73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.766336 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b"} err="failed to get container status \"73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b\": rpc error: code = NotFound desc = could not find container \"73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b\": container with ID starting with 73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b not found: ID does not exist" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.766353 4742 scope.go:117] "RemoveContainer" containerID="78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24" Oct 04 03:20:15 crc kubenswrapper[4742]: E1004 03:20:15.766646 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24\": container with ID starting with 78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24 not found: ID does not exist" containerID="78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.766688 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24"} err="failed to get container status \"78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24\": rpc error: code = NotFound desc = could not find container \"78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24\": container with ID starting with 78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24 not found: ID does not exist" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.766710 4742 scope.go:117] "RemoveContainer" containerID="53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9" Oct 04 03:20:15 crc kubenswrapper[4742]: E1004 03:20:15.767574 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\": container with ID starting with 53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9 not found: ID does not exist" containerID="53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.767637 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9"} err="failed to get container status \"53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\": rpc error: code = NotFound desc = could not find container \"53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\": container with ID starting with 53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9 not found: ID does not exist" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.767660 4742 scope.go:117] "RemoveContainer" containerID="8bf4cc53b1c45b02faf41e5d40191f4fc7ff1959030bfa049955d8e7a31350e7" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.768122 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bf4cc53b1c45b02faf41e5d40191f4fc7ff1959030bfa049955d8e7a31350e7"} err="failed to get container status \"8bf4cc53b1c45b02faf41e5d40191f4fc7ff1959030bfa049955d8e7a31350e7\": rpc error: code = NotFound desc = could not find container \"8bf4cc53b1c45b02faf41e5d40191f4fc7ff1959030bfa049955d8e7a31350e7\": container with ID starting with 8bf4cc53b1c45b02faf41e5d40191f4fc7ff1959030bfa049955d8e7a31350e7 not found: ID does not exist" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.768145 4742 scope.go:117] "RemoveContainer" containerID="e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.768437 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4"} err="failed to get container status \"e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4\": rpc error: code = NotFound desc = could not find container \"e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4\": container with ID starting with e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4 not found: ID does not exist" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.768472 4742 scope.go:117] "RemoveContainer" containerID="0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.769109 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433"} err="failed to get container status \"0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433\": rpc error: code = NotFound desc = could not find container \"0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433\": container with ID starting with 0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433 not found: ID does not exist" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.769147 4742 scope.go:117] "RemoveContainer" containerID="e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.769543 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6"} err="failed to get container status \"e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6\": rpc error: code = NotFound desc = could not find container \"e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6\": container with ID starting with e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6 not found: ID does not exist" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.769585 4742 scope.go:117] "RemoveContainer" containerID="ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.769877 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236"} err="failed to get container status \"ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236\": rpc error: code = NotFound desc = could not find container \"ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236\": container with ID starting with ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236 not found: ID does not exist" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.769905 4742 scope.go:117] "RemoveContainer" containerID="ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.770194 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993"} err="failed to get container status \"ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993\": rpc error: code = NotFound desc = could not find container \"ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993\": container with ID starting with ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993 not found: ID does not exist" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.770228 4742 scope.go:117] "RemoveContainer" containerID="73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.770538 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b"} err="failed to get container status \"73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b\": rpc error: code = NotFound desc = could not find container \"73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b\": container with ID starting with 73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b not found: ID does not exist" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.770580 4742 scope.go:117] "RemoveContainer" containerID="78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.770820 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24"} err="failed to get container status \"78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24\": rpc error: code = NotFound desc = could not find container \"78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24\": container with ID starting with 78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24 not found: ID does not exist" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.770845 4742 scope.go:117] "RemoveContainer" containerID="53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.771081 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9"} err="failed to get container status \"53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\": rpc error: code = NotFound desc = could not find container \"53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\": container with ID starting with 53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9 not found: ID does not exist" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.771118 4742 scope.go:117] "RemoveContainer" containerID="8bf4cc53b1c45b02faf41e5d40191f4fc7ff1959030bfa049955d8e7a31350e7" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.773482 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bf4cc53b1c45b02faf41e5d40191f4fc7ff1959030bfa049955d8e7a31350e7"} err="failed to get container status \"8bf4cc53b1c45b02faf41e5d40191f4fc7ff1959030bfa049955d8e7a31350e7\": rpc error: code = NotFound desc = could not find container \"8bf4cc53b1c45b02faf41e5d40191f4fc7ff1959030bfa049955d8e7a31350e7\": container with ID starting with 8bf4cc53b1c45b02faf41e5d40191f4fc7ff1959030bfa049955d8e7a31350e7 not found: ID does not exist" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.773538 4742 scope.go:117] "RemoveContainer" containerID="e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.773888 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4"} err="failed to get container status \"e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4\": rpc error: code = NotFound desc = could not find container \"e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4\": container with ID starting with e604b30cd0215df9b80de3e3003057b12bdebfc1da32d0c3c3d36db4babe0ea4 not found: ID does not exist" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.773917 4742 scope.go:117] "RemoveContainer" containerID="0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.774299 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433"} err="failed to get container status \"0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433\": rpc error: code = NotFound desc = could not find container \"0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433\": container with ID starting with 0fdae20a37cc8e79e46dbdf95ad434146e5f14e6b71a8eec39ed8e3b69641433 not found: ID does not exist" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.774326 4742 scope.go:117] "RemoveContainer" containerID="e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.774588 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6"} err="failed to get container status \"e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6\": rpc error: code = NotFound desc = could not find container \"e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6\": container with ID starting with e39a2078c4ab3929154ab95be5dbb506c952935492c95c2084922db07a5d6ee6 not found: ID does not exist" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.774609 4742 scope.go:117] "RemoveContainer" containerID="ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.774843 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236"} err="failed to get container status \"ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236\": rpc error: code = NotFound desc = could not find container \"ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236\": container with ID starting with ef094e17130eb455a75544518ce204f4b4e8906e0b65001bff2477d6fe025236 not found: ID does not exist" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.774864 4742 scope.go:117] "RemoveContainer" containerID="ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.775123 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993"} err="failed to get container status \"ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993\": rpc error: code = NotFound desc = could not find container \"ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993\": container with ID starting with ae5dbb64672c7ac0674535e633e9b30f251970995d6e6c549d540a7430938993 not found: ID does not exist" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.775156 4742 scope.go:117] "RemoveContainer" containerID="73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.775506 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b"} err="failed to get container status \"73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b\": rpc error: code = NotFound desc = could not find container \"73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b\": container with ID starting with 73afad36d6ddfcc044b1ceb0d789b80229cbe7e8fcd6b1ef495613a42423057b not found: ID does not exist" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.775537 4742 scope.go:117] "RemoveContainer" containerID="78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.775888 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24"} err="failed to get container status \"78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24\": rpc error: code = NotFound desc = could not find container \"78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24\": container with ID starting with 78ee53cb6a42e187299a90bd25b6c1570799a7117e8f6641015ca75bf839be24 not found: ID does not exist" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.775926 4742 scope.go:117] "RemoveContainer" containerID="53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9" Oct 04 03:20:15 crc kubenswrapper[4742]: I1004 03:20:15.776560 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9"} err="failed to get container status \"53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\": rpc error: code = NotFound desc = could not find container \"53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9\": container with ID starting with 53d4dd4bf44a4287142b72c4da36bbf441b4af6165b0f1f209edf0e18f0c2df9 not found: ID does not exist" Oct 04 03:20:16 crc kubenswrapper[4742]: I1004 03:20:16.551738 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" event={"ID":"e331ee65-e461-4722-8ed4-8af04bd0743c","Type":"ContainerStarted","Data":"c0173064b267a0417dae8e4ff749d0f74856bd76bd79a8fb2feffe02f2104eab"} Oct 04 03:20:16 crc kubenswrapper[4742]: I1004 03:20:16.551778 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" event={"ID":"e331ee65-e461-4722-8ed4-8af04bd0743c","Type":"ContainerStarted","Data":"4d947fe2ee7c6ba120491238e865e45be0f69135def61603b9e546dd8a58305c"} Oct 04 03:20:16 crc kubenswrapper[4742]: I1004 03:20:16.551792 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" event={"ID":"e331ee65-e461-4722-8ed4-8af04bd0743c","Type":"ContainerStarted","Data":"895077deef9ee1c54df927fe5510e54f34dc0ecf6ae60e8db8fd1214ba3de0b9"} Oct 04 03:20:16 crc kubenswrapper[4742]: I1004 03:20:16.551804 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" event={"ID":"e331ee65-e461-4722-8ed4-8af04bd0743c","Type":"ContainerStarted","Data":"be74fa635df9ffcbcddcb098913f97f1bd96730abd3f7258e8bc8c0d56e89f17"} Oct 04 03:20:16 crc kubenswrapper[4742]: I1004 03:20:16.551815 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" event={"ID":"e331ee65-e461-4722-8ed4-8af04bd0743c","Type":"ContainerStarted","Data":"285dd0051876ed337113452f4c2890dc3645811c9324b8ed014aab198de4ece2"} Oct 04 03:20:16 crc kubenswrapper[4742]: I1004 03:20:16.551828 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" event={"ID":"e331ee65-e461-4722-8ed4-8af04bd0743c","Type":"ContainerStarted","Data":"59af779fb0c97afa43e5b063af86de2153587e155a91161c104f0599b12c09f2"} Oct 04 03:20:16 crc kubenswrapper[4742]: I1004 03:20:16.889583 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cfc6e64-ba08-4c44-a18e-3b93ab792a6a" path="/var/lib/kubelet/pods/0cfc6e64-ba08-4c44-a18e-3b93ab792a6a/volumes" Oct 04 03:20:18 crc kubenswrapper[4742]: I1004 03:20:18.567815 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" event={"ID":"e331ee65-e461-4722-8ed4-8af04bd0743c","Type":"ContainerStarted","Data":"c5bbeaa53b135de19d48b028fb45f04bc5658ca91b47335795ecdf539370d5f6"} Oct 04 03:20:21 crc kubenswrapper[4742]: I1004 03:20:21.586436 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" event={"ID":"e331ee65-e461-4722-8ed4-8af04bd0743c","Type":"ContainerStarted","Data":"f820c770984785902923d6f2f02f59c67ae90e826a91dbb67eb996365850df46"} Oct 04 03:20:21 crc kubenswrapper[4742]: I1004 03:20:21.586918 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:21 crc kubenswrapper[4742]: I1004 03:20:21.586996 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:21 crc kubenswrapper[4742]: I1004 03:20:21.587053 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:21 crc kubenswrapper[4742]: I1004 03:20:21.622374 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:21 crc kubenswrapper[4742]: I1004 03:20:21.631005 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:21 crc kubenswrapper[4742]: I1004 03:20:21.640404 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" podStartSLOduration=7.640384589 podStartE2EDuration="7.640384589s" podCreationTimestamp="2025-10-04 03:20:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:20:21.635258496 +0000 UTC m=+587.567595692" watchObservedRunningTime="2025-10-04 03:20:21.640384589 +0000 UTC m=+587.572721765" Oct 04 03:20:28 crc kubenswrapper[4742]: I1004 03:20:28.858346 4742 scope.go:117] "RemoveContainer" containerID="7a1d02fda520e7052329a1d1c9e5516ad70f170819a7d5263271e87b5928ba06" Oct 04 03:20:28 crc kubenswrapper[4742]: E1004 03:20:28.859033 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-nc8m7_openshift-multus(46de52ee-226c-4713-bd7b-961f8f43f635)\"" pod="openshift-multus/multus-nc8m7" podUID="46de52ee-226c-4713-bd7b-961f8f43f635" Oct 04 03:20:38 crc kubenswrapper[4742]: I1004 03:20:38.903951 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6"] Oct 04 03:20:38 crc kubenswrapper[4742]: I1004 03:20:38.910988 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6" Oct 04 03:20:38 crc kubenswrapper[4742]: I1004 03:20:38.912854 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6"] Oct 04 03:20:38 crc kubenswrapper[4742]: I1004 03:20:38.913635 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 04 03:20:39 crc kubenswrapper[4742]: I1004 03:20:39.082690 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlj92\" (UniqueName: \"kubernetes.io/projected/5a2baed0-2265-4564-87fe-ca79418865f6-kube-api-access-tlj92\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6\" (UID: \"5a2baed0-2265-4564-87fe-ca79418865f6\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6" Oct 04 03:20:39 crc kubenswrapper[4742]: I1004 03:20:39.082832 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5a2baed0-2265-4564-87fe-ca79418865f6-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6\" (UID: \"5a2baed0-2265-4564-87fe-ca79418865f6\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6" Oct 04 03:20:39 crc kubenswrapper[4742]: I1004 03:20:39.082858 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5a2baed0-2265-4564-87fe-ca79418865f6-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6\" (UID: \"5a2baed0-2265-4564-87fe-ca79418865f6\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6" Oct 04 03:20:39 crc kubenswrapper[4742]: I1004 03:20:39.184318 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5a2baed0-2265-4564-87fe-ca79418865f6-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6\" (UID: \"5a2baed0-2265-4564-87fe-ca79418865f6\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6" Oct 04 03:20:39 crc kubenswrapper[4742]: I1004 03:20:39.184717 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5a2baed0-2265-4564-87fe-ca79418865f6-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6\" (UID: \"5a2baed0-2265-4564-87fe-ca79418865f6\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6" Oct 04 03:20:39 crc kubenswrapper[4742]: I1004 03:20:39.184858 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlj92\" (UniqueName: \"kubernetes.io/projected/5a2baed0-2265-4564-87fe-ca79418865f6-kube-api-access-tlj92\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6\" (UID: \"5a2baed0-2265-4564-87fe-ca79418865f6\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6" Oct 04 03:20:39 crc kubenswrapper[4742]: I1004 03:20:39.184993 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5a2baed0-2265-4564-87fe-ca79418865f6-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6\" (UID: \"5a2baed0-2265-4564-87fe-ca79418865f6\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6" Oct 04 03:20:39 crc kubenswrapper[4742]: I1004 03:20:39.185206 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5a2baed0-2265-4564-87fe-ca79418865f6-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6\" (UID: \"5a2baed0-2265-4564-87fe-ca79418865f6\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6" Oct 04 03:20:39 crc kubenswrapper[4742]: I1004 03:20:39.204701 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlj92\" (UniqueName: \"kubernetes.io/projected/5a2baed0-2265-4564-87fe-ca79418865f6-kube-api-access-tlj92\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6\" (UID: \"5a2baed0-2265-4564-87fe-ca79418865f6\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6" Oct 04 03:20:39 crc kubenswrapper[4742]: I1004 03:20:39.240783 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6" Oct 04 03:20:39 crc kubenswrapper[4742]: E1004 03:20:39.288885 4742 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6_openshift-marketplace_5a2baed0-2265-4564-87fe-ca79418865f6_0(84946f04ceb7d091a7ab0dbd6512eec58b9e42e8edae2a34cec9b37fa8f42138): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 03:20:39 crc kubenswrapper[4742]: E1004 03:20:39.289354 4742 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6_openshift-marketplace_5a2baed0-2265-4564-87fe-ca79418865f6_0(84946f04ceb7d091a7ab0dbd6512eec58b9e42e8edae2a34cec9b37fa8f42138): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6" Oct 04 03:20:39 crc kubenswrapper[4742]: E1004 03:20:39.289374 4742 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6_openshift-marketplace_5a2baed0-2265-4564-87fe-ca79418865f6_0(84946f04ceb7d091a7ab0dbd6512eec58b9e42e8edae2a34cec9b37fa8f42138): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6" Oct 04 03:20:39 crc kubenswrapper[4742]: E1004 03:20:39.289419 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6_openshift-marketplace(5a2baed0-2265-4564-87fe-ca79418865f6)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6_openshift-marketplace(5a2baed0-2265-4564-87fe-ca79418865f6)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6_openshift-marketplace_5a2baed0-2265-4564-87fe-ca79418865f6_0(84946f04ceb7d091a7ab0dbd6512eec58b9e42e8edae2a34cec9b37fa8f42138): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6" podUID="5a2baed0-2265-4564-87fe-ca79418865f6" Oct 04 03:20:39 crc kubenswrapper[4742]: I1004 03:20:39.690475 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6" Oct 04 03:20:39 crc kubenswrapper[4742]: I1004 03:20:39.690857 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6" Oct 04 03:20:39 crc kubenswrapper[4742]: E1004 03:20:39.717791 4742 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6_openshift-marketplace_5a2baed0-2265-4564-87fe-ca79418865f6_0(8c780c02e602c1ae5aa1816d832ef862818e9b40a0a19fde67d45246bd572070): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 03:20:39 crc kubenswrapper[4742]: E1004 03:20:39.717865 4742 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6_openshift-marketplace_5a2baed0-2265-4564-87fe-ca79418865f6_0(8c780c02e602c1ae5aa1816d832ef862818e9b40a0a19fde67d45246bd572070): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6" Oct 04 03:20:39 crc kubenswrapper[4742]: E1004 03:20:39.717903 4742 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6_openshift-marketplace_5a2baed0-2265-4564-87fe-ca79418865f6_0(8c780c02e602c1ae5aa1816d832ef862818e9b40a0a19fde67d45246bd572070): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6" Oct 04 03:20:39 crc kubenswrapper[4742]: E1004 03:20:39.717977 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6_openshift-marketplace(5a2baed0-2265-4564-87fe-ca79418865f6)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6_openshift-marketplace(5a2baed0-2265-4564-87fe-ca79418865f6)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6_openshift-marketplace_5a2baed0-2265-4564-87fe-ca79418865f6_0(8c780c02e602c1ae5aa1816d832ef862818e9b40a0a19fde67d45246bd572070): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6" podUID="5a2baed0-2265-4564-87fe-ca79418865f6" Oct 04 03:20:41 crc kubenswrapper[4742]: I1004 03:20:41.858809 4742 scope.go:117] "RemoveContainer" containerID="7a1d02fda520e7052329a1d1c9e5516ad70f170819a7d5263271e87b5928ba06" Oct 04 03:20:42 crc kubenswrapper[4742]: I1004 03:20:42.709069 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nc8m7_46de52ee-226c-4713-bd7b-961f8f43f635/kube-multus/2.log" Oct 04 03:20:42 crc kubenswrapper[4742]: I1004 03:20:42.709460 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nc8m7" event={"ID":"46de52ee-226c-4713-bd7b-961f8f43f635","Type":"ContainerStarted","Data":"6bc2e8aad173e861b8eb024659d2422d71d89b64a6129cf10d0d3ada2a3bcc4f"} Oct 04 03:20:44 crc kubenswrapper[4742]: I1004 03:20:44.871729 4742 patch_prober.go:28] interesting pod/machine-config-daemon-qdt7v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:20:44 crc kubenswrapper[4742]: I1004 03:20:44.872567 4742 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:20:45 crc kubenswrapper[4742]: I1004 03:20:45.104675 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ql4ht" Oct 04 03:20:51 crc kubenswrapper[4742]: I1004 03:20:51.857508 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6" Oct 04 03:20:51 crc kubenswrapper[4742]: I1004 03:20:51.859426 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6" Oct 04 03:20:52 crc kubenswrapper[4742]: I1004 03:20:52.139659 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6"] Oct 04 03:20:52 crc kubenswrapper[4742]: I1004 03:20:52.778063 4742 generic.go:334] "Generic (PLEG): container finished" podID="5a2baed0-2265-4564-87fe-ca79418865f6" containerID="2d6eee7186b7840ebe84cb50f6af8a52282b34ea970bb844a6d0d468066f3a6c" exitCode=0 Oct 04 03:20:52 crc kubenswrapper[4742]: I1004 03:20:52.778203 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6" event={"ID":"5a2baed0-2265-4564-87fe-ca79418865f6","Type":"ContainerDied","Data":"2d6eee7186b7840ebe84cb50f6af8a52282b34ea970bb844a6d0d468066f3a6c"} Oct 04 03:20:52 crc kubenswrapper[4742]: I1004 03:20:52.778670 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6" event={"ID":"5a2baed0-2265-4564-87fe-ca79418865f6","Type":"ContainerStarted","Data":"ac627377e2b1334f4632b54c9f18bb7fec32fb91c4f0dfa7113a073b366b6544"} Oct 04 03:20:52 crc kubenswrapper[4742]: I1004 03:20:52.782546 4742 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 03:20:54 crc kubenswrapper[4742]: I1004 03:20:54.807613 4742 generic.go:334] "Generic (PLEG): container finished" podID="5a2baed0-2265-4564-87fe-ca79418865f6" containerID="e13d0de84032787d1c6e5cf9d9a51a3783b68efa44de275ebdf1fd670c7fef85" exitCode=0 Oct 04 03:20:54 crc kubenswrapper[4742]: I1004 03:20:54.807671 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6" event={"ID":"5a2baed0-2265-4564-87fe-ca79418865f6","Type":"ContainerDied","Data":"e13d0de84032787d1c6e5cf9d9a51a3783b68efa44de275ebdf1fd670c7fef85"} Oct 04 03:20:55 crc kubenswrapper[4742]: I1004 03:20:55.819372 4742 generic.go:334] "Generic (PLEG): container finished" podID="5a2baed0-2265-4564-87fe-ca79418865f6" containerID="103bdbe0362ffece9458af2249d169bfe82769a294790b9b5c8a02650494fa76" exitCode=0 Oct 04 03:20:55 crc kubenswrapper[4742]: I1004 03:20:55.819498 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6" event={"ID":"5a2baed0-2265-4564-87fe-ca79418865f6","Type":"ContainerDied","Data":"103bdbe0362ffece9458af2249d169bfe82769a294790b9b5c8a02650494fa76"} Oct 04 03:20:57 crc kubenswrapper[4742]: I1004 03:20:57.120548 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6" Oct 04 03:20:57 crc kubenswrapper[4742]: I1004 03:20:57.147229 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tlj92\" (UniqueName: \"kubernetes.io/projected/5a2baed0-2265-4564-87fe-ca79418865f6-kube-api-access-tlj92\") pod \"5a2baed0-2265-4564-87fe-ca79418865f6\" (UID: \"5a2baed0-2265-4564-87fe-ca79418865f6\") " Oct 04 03:20:57 crc kubenswrapper[4742]: I1004 03:20:57.147901 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5a2baed0-2265-4564-87fe-ca79418865f6-bundle\") pod \"5a2baed0-2265-4564-87fe-ca79418865f6\" (UID: \"5a2baed0-2265-4564-87fe-ca79418865f6\") " Oct 04 03:20:57 crc kubenswrapper[4742]: I1004 03:20:57.147927 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5a2baed0-2265-4564-87fe-ca79418865f6-util\") pod \"5a2baed0-2265-4564-87fe-ca79418865f6\" (UID: \"5a2baed0-2265-4564-87fe-ca79418865f6\") " Oct 04 03:20:57 crc kubenswrapper[4742]: I1004 03:20:57.149142 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a2baed0-2265-4564-87fe-ca79418865f6-bundle" (OuterVolumeSpecName: "bundle") pod "5a2baed0-2265-4564-87fe-ca79418865f6" (UID: "5a2baed0-2265-4564-87fe-ca79418865f6"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:20:57 crc kubenswrapper[4742]: I1004 03:20:57.157611 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a2baed0-2265-4564-87fe-ca79418865f6-kube-api-access-tlj92" (OuterVolumeSpecName: "kube-api-access-tlj92") pod "5a2baed0-2265-4564-87fe-ca79418865f6" (UID: "5a2baed0-2265-4564-87fe-ca79418865f6"). InnerVolumeSpecName "kube-api-access-tlj92". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:20:57 crc kubenswrapper[4742]: I1004 03:20:57.168875 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a2baed0-2265-4564-87fe-ca79418865f6-util" (OuterVolumeSpecName: "util") pod "5a2baed0-2265-4564-87fe-ca79418865f6" (UID: "5a2baed0-2265-4564-87fe-ca79418865f6"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:20:57 crc kubenswrapper[4742]: I1004 03:20:57.249825 4742 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5a2baed0-2265-4564-87fe-ca79418865f6-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:57 crc kubenswrapper[4742]: I1004 03:20:57.249872 4742 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5a2baed0-2265-4564-87fe-ca79418865f6-util\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:57 crc kubenswrapper[4742]: I1004 03:20:57.249890 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tlj92\" (UniqueName: \"kubernetes.io/projected/5a2baed0-2265-4564-87fe-ca79418865f6-kube-api-access-tlj92\") on node \"crc\" DevicePath \"\"" Oct 04 03:20:57 crc kubenswrapper[4742]: I1004 03:20:57.832960 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6" event={"ID":"5a2baed0-2265-4564-87fe-ca79418865f6","Type":"ContainerDied","Data":"ac627377e2b1334f4632b54c9f18bb7fec32fb91c4f0dfa7113a073b366b6544"} Oct 04 03:20:57 crc kubenswrapper[4742]: I1004 03:20:57.833026 4742 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac627377e2b1334f4632b54c9f18bb7fec32fb91c4f0dfa7113a073b366b6544" Oct 04 03:20:57 crc kubenswrapper[4742]: I1004 03:20:57.833003 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6" Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.046227 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-c97f8c589-xphrs"] Oct 04 03:21:07 crc kubenswrapper[4742]: E1004 03:21:07.047483 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a2baed0-2265-4564-87fe-ca79418865f6" containerName="extract" Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.047501 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a2baed0-2265-4564-87fe-ca79418865f6" containerName="extract" Oct 04 03:21:07 crc kubenswrapper[4742]: E1004 03:21:07.047513 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a2baed0-2265-4564-87fe-ca79418865f6" containerName="util" Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.047519 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a2baed0-2265-4564-87fe-ca79418865f6" containerName="util" Oct 04 03:21:07 crc kubenswrapper[4742]: E1004 03:21:07.047529 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a2baed0-2265-4564-87fe-ca79418865f6" containerName="pull" Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.047535 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a2baed0-2265-4564-87fe-ca79418865f6" containerName="pull" Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.047660 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a2baed0-2265-4564-87fe-ca79418865f6" containerName="extract" Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.048191 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-c97f8c589-xphrs" Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.051114 4742 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.052314 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.052913 4742 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.054493 4742 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-q8blk" Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.064027 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-c97f8c589-xphrs"] Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.066058 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.075187 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9bcd94cc-6df7-4f9c-be06-2234b0d99d72-webhook-cert\") pod \"metallb-operator-controller-manager-c97f8c589-xphrs\" (UID: \"9bcd94cc-6df7-4f9c-be06-2234b0d99d72\") " pod="metallb-system/metallb-operator-controller-manager-c97f8c589-xphrs" Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.075445 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjqzg\" (UniqueName: \"kubernetes.io/projected/9bcd94cc-6df7-4f9c-be06-2234b0d99d72-kube-api-access-hjqzg\") pod \"metallb-operator-controller-manager-c97f8c589-xphrs\" (UID: \"9bcd94cc-6df7-4f9c-be06-2234b0d99d72\") " pod="metallb-system/metallb-operator-controller-manager-c97f8c589-xphrs" Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.075564 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9bcd94cc-6df7-4f9c-be06-2234b0d99d72-apiservice-cert\") pod \"metallb-operator-controller-manager-c97f8c589-xphrs\" (UID: \"9bcd94cc-6df7-4f9c-be06-2234b0d99d72\") " pod="metallb-system/metallb-operator-controller-manager-c97f8c589-xphrs" Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.177097 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjqzg\" (UniqueName: \"kubernetes.io/projected/9bcd94cc-6df7-4f9c-be06-2234b0d99d72-kube-api-access-hjqzg\") pod \"metallb-operator-controller-manager-c97f8c589-xphrs\" (UID: \"9bcd94cc-6df7-4f9c-be06-2234b0d99d72\") " pod="metallb-system/metallb-operator-controller-manager-c97f8c589-xphrs" Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.177511 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9bcd94cc-6df7-4f9c-be06-2234b0d99d72-apiservice-cert\") pod \"metallb-operator-controller-manager-c97f8c589-xphrs\" (UID: \"9bcd94cc-6df7-4f9c-be06-2234b0d99d72\") " pod="metallb-system/metallb-operator-controller-manager-c97f8c589-xphrs" Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.177623 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9bcd94cc-6df7-4f9c-be06-2234b0d99d72-webhook-cert\") pod \"metallb-operator-controller-manager-c97f8c589-xphrs\" (UID: \"9bcd94cc-6df7-4f9c-be06-2234b0d99d72\") " pod="metallb-system/metallb-operator-controller-manager-c97f8c589-xphrs" Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.186107 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9bcd94cc-6df7-4f9c-be06-2234b0d99d72-apiservice-cert\") pod \"metallb-operator-controller-manager-c97f8c589-xphrs\" (UID: \"9bcd94cc-6df7-4f9c-be06-2234b0d99d72\") " pod="metallb-system/metallb-operator-controller-manager-c97f8c589-xphrs" Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.189524 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9bcd94cc-6df7-4f9c-be06-2234b0d99d72-webhook-cert\") pod \"metallb-operator-controller-manager-c97f8c589-xphrs\" (UID: \"9bcd94cc-6df7-4f9c-be06-2234b0d99d72\") " pod="metallb-system/metallb-operator-controller-manager-c97f8c589-xphrs" Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.222910 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjqzg\" (UniqueName: \"kubernetes.io/projected/9bcd94cc-6df7-4f9c-be06-2234b0d99d72-kube-api-access-hjqzg\") pod \"metallb-operator-controller-manager-c97f8c589-xphrs\" (UID: \"9bcd94cc-6df7-4f9c-be06-2234b0d99d72\") " pod="metallb-system/metallb-operator-controller-manager-c97f8c589-xphrs" Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.367897 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-c97f8c589-xphrs" Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.370995 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-75697b4fb9-tnl24"] Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.371794 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-75697b4fb9-tnl24" Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.382581 4742 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.382790 4742 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-ljvh7" Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.385904 4742 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.445327 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-75697b4fb9-tnl24"] Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.481883 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e401237a-d63b-4e6a-8efc-58b9ed5a6846-webhook-cert\") pod \"metallb-operator-webhook-server-75697b4fb9-tnl24\" (UID: \"e401237a-d63b-4e6a-8efc-58b9ed5a6846\") " pod="metallb-system/metallb-operator-webhook-server-75697b4fb9-tnl24" Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.482412 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s86dp\" (UniqueName: \"kubernetes.io/projected/e401237a-d63b-4e6a-8efc-58b9ed5a6846-kube-api-access-s86dp\") pod \"metallb-operator-webhook-server-75697b4fb9-tnl24\" (UID: \"e401237a-d63b-4e6a-8efc-58b9ed5a6846\") " pod="metallb-system/metallb-operator-webhook-server-75697b4fb9-tnl24" Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.482487 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e401237a-d63b-4e6a-8efc-58b9ed5a6846-apiservice-cert\") pod \"metallb-operator-webhook-server-75697b4fb9-tnl24\" (UID: \"e401237a-d63b-4e6a-8efc-58b9ed5a6846\") " pod="metallb-system/metallb-operator-webhook-server-75697b4fb9-tnl24" Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.587333 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s86dp\" (UniqueName: \"kubernetes.io/projected/e401237a-d63b-4e6a-8efc-58b9ed5a6846-kube-api-access-s86dp\") pod \"metallb-operator-webhook-server-75697b4fb9-tnl24\" (UID: \"e401237a-d63b-4e6a-8efc-58b9ed5a6846\") " pod="metallb-system/metallb-operator-webhook-server-75697b4fb9-tnl24" Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.587444 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e401237a-d63b-4e6a-8efc-58b9ed5a6846-apiservice-cert\") pod \"metallb-operator-webhook-server-75697b4fb9-tnl24\" (UID: \"e401237a-d63b-4e6a-8efc-58b9ed5a6846\") " pod="metallb-system/metallb-operator-webhook-server-75697b4fb9-tnl24" Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.587507 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e401237a-d63b-4e6a-8efc-58b9ed5a6846-webhook-cert\") pod \"metallb-operator-webhook-server-75697b4fb9-tnl24\" (UID: \"e401237a-d63b-4e6a-8efc-58b9ed5a6846\") " pod="metallb-system/metallb-operator-webhook-server-75697b4fb9-tnl24" Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.593292 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e401237a-d63b-4e6a-8efc-58b9ed5a6846-webhook-cert\") pod \"metallb-operator-webhook-server-75697b4fb9-tnl24\" (UID: \"e401237a-d63b-4e6a-8efc-58b9ed5a6846\") " pod="metallb-system/metallb-operator-webhook-server-75697b4fb9-tnl24" Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.593336 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e401237a-d63b-4e6a-8efc-58b9ed5a6846-apiservice-cert\") pod \"metallb-operator-webhook-server-75697b4fb9-tnl24\" (UID: \"e401237a-d63b-4e6a-8efc-58b9ed5a6846\") " pod="metallb-system/metallb-operator-webhook-server-75697b4fb9-tnl24" Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.609163 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s86dp\" (UniqueName: \"kubernetes.io/projected/e401237a-d63b-4e6a-8efc-58b9ed5a6846-kube-api-access-s86dp\") pod \"metallb-operator-webhook-server-75697b4fb9-tnl24\" (UID: \"e401237a-d63b-4e6a-8efc-58b9ed5a6846\") " pod="metallb-system/metallb-operator-webhook-server-75697b4fb9-tnl24" Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.717387 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-75697b4fb9-tnl24" Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.825850 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-c97f8c589-xphrs"] Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.915692 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-c97f8c589-xphrs" event={"ID":"9bcd94cc-6df7-4f9c-be06-2234b0d99d72","Type":"ContainerStarted","Data":"a1fa41b84100a5124ef5fabe8750e88b502f02bc1c1ad7a164297dae2bd3def9"} Oct 04 03:21:07 crc kubenswrapper[4742]: I1004 03:21:07.983784 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-75697b4fb9-tnl24"] Oct 04 03:21:07 crc kubenswrapper[4742]: W1004 03:21:07.985643 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode401237a_d63b_4e6a_8efc_58b9ed5a6846.slice/crio-3e7e955b183f5d07318e12425bce4da1bdc36938131305d1331c13a5667f2826 WatchSource:0}: Error finding container 3e7e955b183f5d07318e12425bce4da1bdc36938131305d1331c13a5667f2826: Status 404 returned error can't find the container with id 3e7e955b183f5d07318e12425bce4da1bdc36938131305d1331c13a5667f2826 Oct 04 03:21:08 crc kubenswrapper[4742]: I1004 03:21:08.921509 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-75697b4fb9-tnl24" event={"ID":"e401237a-d63b-4e6a-8efc-58b9ed5a6846","Type":"ContainerStarted","Data":"3e7e955b183f5d07318e12425bce4da1bdc36938131305d1331c13a5667f2826"} Oct 04 03:21:12 crc kubenswrapper[4742]: I1004 03:21:12.955442 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-c97f8c589-xphrs" event={"ID":"9bcd94cc-6df7-4f9c-be06-2234b0d99d72","Type":"ContainerStarted","Data":"af12b5e917e859ea3833f3634a895924d94edb88b94c4b59ba7f6e3b269cff9e"} Oct 04 03:21:12 crc kubenswrapper[4742]: I1004 03:21:12.957229 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-c97f8c589-xphrs" Oct 04 03:21:12 crc kubenswrapper[4742]: I1004 03:21:12.959860 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-75697b4fb9-tnl24" event={"ID":"e401237a-d63b-4e6a-8efc-58b9ed5a6846","Type":"ContainerStarted","Data":"6a3d40b933d2c964ee69192b3771ecacc4e5204cd5f9e357f5f3b172f7e0eff8"} Oct 04 03:21:12 crc kubenswrapper[4742]: I1004 03:21:12.961677 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-75697b4fb9-tnl24" Oct 04 03:21:12 crc kubenswrapper[4742]: I1004 03:21:12.985747 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-c97f8c589-xphrs" podStartSLOduration=2.900310704 podStartE2EDuration="5.98566935s" podCreationTimestamp="2025-10-04 03:21:07 +0000 UTC" firstStartedPulling="2025-10-04 03:21:07.846039925 +0000 UTC m=+633.778377101" lastFinishedPulling="2025-10-04 03:21:10.931398571 +0000 UTC m=+636.863735747" observedRunningTime="2025-10-04 03:21:12.978484715 +0000 UTC m=+638.910821891" watchObservedRunningTime="2025-10-04 03:21:12.98566935 +0000 UTC m=+638.918006536" Oct 04 03:21:13 crc kubenswrapper[4742]: I1004 03:21:13.010237 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-75697b4fb9-tnl24" podStartSLOduration=1.5172882890000001 podStartE2EDuration="6.010198942s" podCreationTimestamp="2025-10-04 03:21:07 +0000 UTC" firstStartedPulling="2025-10-04 03:21:07.98867343 +0000 UTC m=+633.921010606" lastFinishedPulling="2025-10-04 03:21:12.481584083 +0000 UTC m=+638.413921259" observedRunningTime="2025-10-04 03:21:13.006446895 +0000 UTC m=+638.938784101" watchObservedRunningTime="2025-10-04 03:21:13.010198942 +0000 UTC m=+638.942536158" Oct 04 03:21:14 crc kubenswrapper[4742]: I1004 03:21:14.871679 4742 patch_prober.go:28] interesting pod/machine-config-daemon-qdt7v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:21:14 crc kubenswrapper[4742]: I1004 03:21:14.871757 4742 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:21:14 crc kubenswrapper[4742]: I1004 03:21:14.871812 4742 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" Oct 04 03:21:14 crc kubenswrapper[4742]: I1004 03:21:14.872496 4742 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"716e86aaee18fb777354d280f34d7ae9aea9909aae176a0c4c726413f1e1cb06"} pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 03:21:14 crc kubenswrapper[4742]: I1004 03:21:14.872575 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerName="machine-config-daemon" containerID="cri-o://716e86aaee18fb777354d280f34d7ae9aea9909aae176a0c4c726413f1e1cb06" gracePeriod=600 Oct 04 03:21:15 crc kubenswrapper[4742]: I1004 03:21:15.976501 4742 generic.go:334] "Generic (PLEG): container finished" podID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerID="716e86aaee18fb777354d280f34d7ae9aea9909aae176a0c4c726413f1e1cb06" exitCode=0 Oct 04 03:21:15 crc kubenswrapper[4742]: I1004 03:21:15.976578 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" event={"ID":"989ea247-7ac9-4ade-8dcb-2632fef4f83e","Type":"ContainerDied","Data":"716e86aaee18fb777354d280f34d7ae9aea9909aae176a0c4c726413f1e1cb06"} Oct 04 03:21:15 crc kubenswrapper[4742]: I1004 03:21:15.977120 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" event={"ID":"989ea247-7ac9-4ade-8dcb-2632fef4f83e","Type":"ContainerStarted","Data":"ffcec0a1f833323e5678d3dde9b3cc1a6ca05ee38cfe04eae68bdf431a24a5f2"} Oct 04 03:21:15 crc kubenswrapper[4742]: I1004 03:21:15.977162 4742 scope.go:117] "RemoveContainer" containerID="265d94a26e57f0eae37e63999ff3e9ea88f239556449abe4f5a9ca7688a659d5" Oct 04 03:21:27 crc kubenswrapper[4742]: I1004 03:21:27.723350 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-75697b4fb9-tnl24" Oct 04 03:21:47 crc kubenswrapper[4742]: I1004 03:21:47.371631 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-c97f8c589-xphrs" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.051598 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-b4g9q"] Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.052532 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-b4g9q" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.054846 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-zc8db"] Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.056204 4742 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.056906 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-zc8db" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.058449 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.058612 4742 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.059444 4742 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-vkx9x" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.118321 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-b4g9q"] Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.125829 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-r985t"] Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.126750 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-r985t" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.129092 4742 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.129316 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.129599 4742 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.129862 4742 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-dmnwz" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.135051 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-cdc8r"] Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.135856 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-cdc8r" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.138056 4742 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.148084 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-cdc8r"] Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.228760 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b43429d7-ac32-4ab9-91e1-1ac6010312ec-memberlist\") pod \"speaker-r985t\" (UID: \"b43429d7-ac32-4ab9-91e1-1ac6010312ec\") " pod="metallb-system/speaker-r985t" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.228797 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b43429d7-ac32-4ab9-91e1-1ac6010312ec-metrics-certs\") pod \"speaker-r985t\" (UID: \"b43429d7-ac32-4ab9-91e1-1ac6010312ec\") " pod="metallb-system/speaker-r985t" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.228815 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9c0b8053-d605-4419-81fa-e7f17c45e068-cert\") pod \"controller-68d546b9d8-cdc8r\" (UID: \"9c0b8053-d605-4419-81fa-e7f17c45e068\") " pod="metallb-system/controller-68d546b9d8-cdc8r" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.228834 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9c0b8053-d605-4419-81fa-e7f17c45e068-metrics-certs\") pod \"controller-68d546b9d8-cdc8r\" (UID: \"9c0b8053-d605-4419-81fa-e7f17c45e068\") " pod="metallb-system/controller-68d546b9d8-cdc8r" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.228851 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/37f0f686-5a17-4d4a-87d8-0f1d3410530f-frr-sockets\") pod \"frr-k8s-zc8db\" (UID: \"37f0f686-5a17-4d4a-87d8-0f1d3410530f\") " pod="metallb-system/frr-k8s-zc8db" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.228865 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/37f0f686-5a17-4d4a-87d8-0f1d3410530f-metrics\") pod \"frr-k8s-zc8db\" (UID: \"37f0f686-5a17-4d4a-87d8-0f1d3410530f\") " pod="metallb-system/frr-k8s-zc8db" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.228937 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/b43429d7-ac32-4ab9-91e1-1ac6010312ec-metallb-excludel2\") pod \"speaker-r985t\" (UID: \"b43429d7-ac32-4ab9-91e1-1ac6010312ec\") " pod="metallb-system/speaker-r985t" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.228953 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z798f\" (UniqueName: \"kubernetes.io/projected/b43429d7-ac32-4ab9-91e1-1ac6010312ec-kube-api-access-z798f\") pod \"speaker-r985t\" (UID: \"b43429d7-ac32-4ab9-91e1-1ac6010312ec\") " pod="metallb-system/speaker-r985t" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.228973 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/37f0f686-5a17-4d4a-87d8-0f1d3410530f-metrics-certs\") pod \"frr-k8s-zc8db\" (UID: \"37f0f686-5a17-4d4a-87d8-0f1d3410530f\") " pod="metallb-system/frr-k8s-zc8db" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.229000 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/37f0f686-5a17-4d4a-87d8-0f1d3410530f-frr-conf\") pod \"frr-k8s-zc8db\" (UID: \"37f0f686-5a17-4d4a-87d8-0f1d3410530f\") " pod="metallb-system/frr-k8s-zc8db" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.229017 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6lxr\" (UniqueName: \"kubernetes.io/projected/9c0b8053-d605-4419-81fa-e7f17c45e068-kube-api-access-d6lxr\") pod \"controller-68d546b9d8-cdc8r\" (UID: \"9c0b8053-d605-4419-81fa-e7f17c45e068\") " pod="metallb-system/controller-68d546b9d8-cdc8r" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.229157 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c32d6520-f0f0-4ba4-8317-0208036377db-cert\") pod \"frr-k8s-webhook-server-64bf5d555-b4g9q\" (UID: \"c32d6520-f0f0-4ba4-8317-0208036377db\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-b4g9q" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.229357 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/37f0f686-5a17-4d4a-87d8-0f1d3410530f-frr-startup\") pod \"frr-k8s-zc8db\" (UID: \"37f0f686-5a17-4d4a-87d8-0f1d3410530f\") " pod="metallb-system/frr-k8s-zc8db" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.229420 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/37f0f686-5a17-4d4a-87d8-0f1d3410530f-reloader\") pod \"frr-k8s-zc8db\" (UID: \"37f0f686-5a17-4d4a-87d8-0f1d3410530f\") " pod="metallb-system/frr-k8s-zc8db" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.229478 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msf6r\" (UniqueName: \"kubernetes.io/projected/37f0f686-5a17-4d4a-87d8-0f1d3410530f-kube-api-access-msf6r\") pod \"frr-k8s-zc8db\" (UID: \"37f0f686-5a17-4d4a-87d8-0f1d3410530f\") " pod="metallb-system/frr-k8s-zc8db" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.229557 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fggt7\" (UniqueName: \"kubernetes.io/projected/c32d6520-f0f0-4ba4-8317-0208036377db-kube-api-access-fggt7\") pod \"frr-k8s-webhook-server-64bf5d555-b4g9q\" (UID: \"c32d6520-f0f0-4ba4-8317-0208036377db\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-b4g9q" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.331351 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b43429d7-ac32-4ab9-91e1-1ac6010312ec-metrics-certs\") pod \"speaker-r985t\" (UID: \"b43429d7-ac32-4ab9-91e1-1ac6010312ec\") " pod="metallb-system/speaker-r985t" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.331631 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b43429d7-ac32-4ab9-91e1-1ac6010312ec-memberlist\") pod \"speaker-r985t\" (UID: \"b43429d7-ac32-4ab9-91e1-1ac6010312ec\") " pod="metallb-system/speaker-r985t" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.331713 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9c0b8053-d605-4419-81fa-e7f17c45e068-cert\") pod \"controller-68d546b9d8-cdc8r\" (UID: \"9c0b8053-d605-4419-81fa-e7f17c45e068\") " pod="metallb-system/controller-68d546b9d8-cdc8r" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.331796 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9c0b8053-d605-4419-81fa-e7f17c45e068-metrics-certs\") pod \"controller-68d546b9d8-cdc8r\" (UID: \"9c0b8053-d605-4419-81fa-e7f17c45e068\") " pod="metallb-system/controller-68d546b9d8-cdc8r" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.331872 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/37f0f686-5a17-4d4a-87d8-0f1d3410530f-frr-sockets\") pod \"frr-k8s-zc8db\" (UID: \"37f0f686-5a17-4d4a-87d8-0f1d3410530f\") " pod="metallb-system/frr-k8s-zc8db" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.331950 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/37f0f686-5a17-4d4a-87d8-0f1d3410530f-metrics\") pod \"frr-k8s-zc8db\" (UID: \"37f0f686-5a17-4d4a-87d8-0f1d3410530f\") " pod="metallb-system/frr-k8s-zc8db" Oct 04 03:21:48 crc kubenswrapper[4742]: E1004 03:21:48.331734 4742 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 04 03:21:48 crc kubenswrapper[4742]: E1004 03:21:48.332121 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b43429d7-ac32-4ab9-91e1-1ac6010312ec-memberlist podName:b43429d7-ac32-4ab9-91e1-1ac6010312ec nodeName:}" failed. No retries permitted until 2025-10-04 03:21:48.832089796 +0000 UTC m=+674.764426972 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/b43429d7-ac32-4ab9-91e1-1ac6010312ec-memberlist") pod "speaker-r985t" (UID: "b43429d7-ac32-4ab9-91e1-1ac6010312ec") : secret "metallb-memberlist" not found Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.332156 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/b43429d7-ac32-4ab9-91e1-1ac6010312ec-metallb-excludel2\") pod \"speaker-r985t\" (UID: \"b43429d7-ac32-4ab9-91e1-1ac6010312ec\") " pod="metallb-system/speaker-r985t" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.332191 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z798f\" (UniqueName: \"kubernetes.io/projected/b43429d7-ac32-4ab9-91e1-1ac6010312ec-kube-api-access-z798f\") pod \"speaker-r985t\" (UID: \"b43429d7-ac32-4ab9-91e1-1ac6010312ec\") " pod="metallb-system/speaker-r985t" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.332238 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/37f0f686-5a17-4d4a-87d8-0f1d3410530f-metrics-certs\") pod \"frr-k8s-zc8db\" (UID: \"37f0f686-5a17-4d4a-87d8-0f1d3410530f\") " pod="metallb-system/frr-k8s-zc8db" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.332243 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/37f0f686-5a17-4d4a-87d8-0f1d3410530f-frr-sockets\") pod \"frr-k8s-zc8db\" (UID: \"37f0f686-5a17-4d4a-87d8-0f1d3410530f\") " pod="metallb-system/frr-k8s-zc8db" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.332360 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/37f0f686-5a17-4d4a-87d8-0f1d3410530f-frr-conf\") pod \"frr-k8s-zc8db\" (UID: \"37f0f686-5a17-4d4a-87d8-0f1d3410530f\") " pod="metallb-system/frr-k8s-zc8db" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.332390 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6lxr\" (UniqueName: \"kubernetes.io/projected/9c0b8053-d605-4419-81fa-e7f17c45e068-kube-api-access-d6lxr\") pod \"controller-68d546b9d8-cdc8r\" (UID: \"9c0b8053-d605-4419-81fa-e7f17c45e068\") " pod="metallb-system/controller-68d546b9d8-cdc8r" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.332416 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c32d6520-f0f0-4ba4-8317-0208036377db-cert\") pod \"frr-k8s-webhook-server-64bf5d555-b4g9q\" (UID: \"c32d6520-f0f0-4ba4-8317-0208036377db\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-b4g9q" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.332468 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/37f0f686-5a17-4d4a-87d8-0f1d3410530f-frr-startup\") pod \"frr-k8s-zc8db\" (UID: \"37f0f686-5a17-4d4a-87d8-0f1d3410530f\") " pod="metallb-system/frr-k8s-zc8db" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.332483 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/37f0f686-5a17-4d4a-87d8-0f1d3410530f-reloader\") pod \"frr-k8s-zc8db\" (UID: \"37f0f686-5a17-4d4a-87d8-0f1d3410530f\") " pod="metallb-system/frr-k8s-zc8db" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.332509 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msf6r\" (UniqueName: \"kubernetes.io/projected/37f0f686-5a17-4d4a-87d8-0f1d3410530f-kube-api-access-msf6r\") pod \"frr-k8s-zc8db\" (UID: \"37f0f686-5a17-4d4a-87d8-0f1d3410530f\") " pod="metallb-system/frr-k8s-zc8db" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.332549 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fggt7\" (UniqueName: \"kubernetes.io/projected/c32d6520-f0f0-4ba4-8317-0208036377db-kube-api-access-fggt7\") pod \"frr-k8s-webhook-server-64bf5d555-b4g9q\" (UID: \"c32d6520-f0f0-4ba4-8317-0208036377db\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-b4g9q" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.332778 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/37f0f686-5a17-4d4a-87d8-0f1d3410530f-metrics\") pod \"frr-k8s-zc8db\" (UID: \"37f0f686-5a17-4d4a-87d8-0f1d3410530f\") " pod="metallb-system/frr-k8s-zc8db" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.333054 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/37f0f686-5a17-4d4a-87d8-0f1d3410530f-reloader\") pod \"frr-k8s-zc8db\" (UID: \"37f0f686-5a17-4d4a-87d8-0f1d3410530f\") " pod="metallb-system/frr-k8s-zc8db" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.333055 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/b43429d7-ac32-4ab9-91e1-1ac6010312ec-metallb-excludel2\") pod \"speaker-r985t\" (UID: \"b43429d7-ac32-4ab9-91e1-1ac6010312ec\") " pod="metallb-system/speaker-r985t" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.333062 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/37f0f686-5a17-4d4a-87d8-0f1d3410530f-frr-conf\") pod \"frr-k8s-zc8db\" (UID: \"37f0f686-5a17-4d4a-87d8-0f1d3410530f\") " pod="metallb-system/frr-k8s-zc8db" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.333360 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/37f0f686-5a17-4d4a-87d8-0f1d3410530f-frr-startup\") pod \"frr-k8s-zc8db\" (UID: \"37f0f686-5a17-4d4a-87d8-0f1d3410530f\") " pod="metallb-system/frr-k8s-zc8db" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.336444 4742 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.338376 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9c0b8053-d605-4419-81fa-e7f17c45e068-metrics-certs\") pod \"controller-68d546b9d8-cdc8r\" (UID: \"9c0b8053-d605-4419-81fa-e7f17c45e068\") " pod="metallb-system/controller-68d546b9d8-cdc8r" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.338398 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b43429d7-ac32-4ab9-91e1-1ac6010312ec-metrics-certs\") pod \"speaker-r985t\" (UID: \"b43429d7-ac32-4ab9-91e1-1ac6010312ec\") " pod="metallb-system/speaker-r985t" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.342804 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/37f0f686-5a17-4d4a-87d8-0f1d3410530f-metrics-certs\") pod \"frr-k8s-zc8db\" (UID: \"37f0f686-5a17-4d4a-87d8-0f1d3410530f\") " pod="metallb-system/frr-k8s-zc8db" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.345213 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9c0b8053-d605-4419-81fa-e7f17c45e068-cert\") pod \"controller-68d546b9d8-cdc8r\" (UID: \"9c0b8053-d605-4419-81fa-e7f17c45e068\") " pod="metallb-system/controller-68d546b9d8-cdc8r" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.349916 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c32d6520-f0f0-4ba4-8317-0208036377db-cert\") pod \"frr-k8s-webhook-server-64bf5d555-b4g9q\" (UID: \"c32d6520-f0f0-4ba4-8317-0208036377db\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-b4g9q" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.354170 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6lxr\" (UniqueName: \"kubernetes.io/projected/9c0b8053-d605-4419-81fa-e7f17c45e068-kube-api-access-d6lxr\") pod \"controller-68d546b9d8-cdc8r\" (UID: \"9c0b8053-d605-4419-81fa-e7f17c45e068\") " pod="metallb-system/controller-68d546b9d8-cdc8r" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.355220 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msf6r\" (UniqueName: \"kubernetes.io/projected/37f0f686-5a17-4d4a-87d8-0f1d3410530f-kube-api-access-msf6r\") pod \"frr-k8s-zc8db\" (UID: \"37f0f686-5a17-4d4a-87d8-0f1d3410530f\") " pod="metallb-system/frr-k8s-zc8db" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.355518 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z798f\" (UniqueName: \"kubernetes.io/projected/b43429d7-ac32-4ab9-91e1-1ac6010312ec-kube-api-access-z798f\") pod \"speaker-r985t\" (UID: \"b43429d7-ac32-4ab9-91e1-1ac6010312ec\") " pod="metallb-system/speaker-r985t" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.358601 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fggt7\" (UniqueName: \"kubernetes.io/projected/c32d6520-f0f0-4ba4-8317-0208036377db-kube-api-access-fggt7\") pod \"frr-k8s-webhook-server-64bf5d555-b4g9q\" (UID: \"c32d6520-f0f0-4ba4-8317-0208036377db\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-b4g9q" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.372447 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-b4g9q" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.379660 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-zc8db" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.449590 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-cdc8r" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.635181 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-b4g9q"] Oct 04 03:21:48 crc kubenswrapper[4742]: W1004 03:21:48.642594 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc32d6520_f0f0_4ba4_8317_0208036377db.slice/crio-5ed0e9e10f4527ae49679283d8aa7913ac2d732e03d7242790c3420858015998 WatchSource:0}: Error finding container 5ed0e9e10f4527ae49679283d8aa7913ac2d732e03d7242790c3420858015998: Status 404 returned error can't find the container with id 5ed0e9e10f4527ae49679283d8aa7913ac2d732e03d7242790c3420858015998 Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.671280 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-cdc8r"] Oct 04 03:21:48 crc kubenswrapper[4742]: W1004 03:21:48.675788 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9c0b8053_d605_4419_81fa_e7f17c45e068.slice/crio-d23076c9dc83b2d98cc50b8cf1cf194f69dc27e5c78550d1fbef7cdd23135bde WatchSource:0}: Error finding container d23076c9dc83b2d98cc50b8cf1cf194f69dc27e5c78550d1fbef7cdd23135bde: Status 404 returned error can't find the container with id d23076c9dc83b2d98cc50b8cf1cf194f69dc27e5c78550d1fbef7cdd23135bde Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.838960 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b43429d7-ac32-4ab9-91e1-1ac6010312ec-memberlist\") pod \"speaker-r985t\" (UID: \"b43429d7-ac32-4ab9-91e1-1ac6010312ec\") " pod="metallb-system/speaker-r985t" Oct 04 03:21:48 crc kubenswrapper[4742]: I1004 03:21:48.843892 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b43429d7-ac32-4ab9-91e1-1ac6010312ec-memberlist\") pod \"speaker-r985t\" (UID: \"b43429d7-ac32-4ab9-91e1-1ac6010312ec\") " pod="metallb-system/speaker-r985t" Oct 04 03:21:49 crc kubenswrapper[4742]: I1004 03:21:49.041450 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-r985t" Oct 04 03:21:49 crc kubenswrapper[4742]: W1004 03:21:49.058065 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb43429d7_ac32_4ab9_91e1_1ac6010312ec.slice/crio-4379f428fd477a350686aad7aa1f6e937a72fb44e6d5e3b915292d820efe0ce7 WatchSource:0}: Error finding container 4379f428fd477a350686aad7aa1f6e937a72fb44e6d5e3b915292d820efe0ce7: Status 404 returned error can't find the container with id 4379f428fd477a350686aad7aa1f6e937a72fb44e6d5e3b915292d820efe0ce7 Oct 04 03:21:49 crc kubenswrapper[4742]: I1004 03:21:49.148385 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zc8db" event={"ID":"37f0f686-5a17-4d4a-87d8-0f1d3410530f","Type":"ContainerStarted","Data":"16e714d5e80f2aa32caa0ff6353ff2d029f1b2cee31066ceec20c2b94a1fdd4a"} Oct 04 03:21:49 crc kubenswrapper[4742]: I1004 03:21:49.151199 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-cdc8r" event={"ID":"9c0b8053-d605-4419-81fa-e7f17c45e068","Type":"ContainerStarted","Data":"cb5793398bfa37bf79a5811f00602e8be87248857f0fcdaa954d5109fa3f1301"} Oct 04 03:21:49 crc kubenswrapper[4742]: I1004 03:21:49.151224 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-cdc8r" event={"ID":"9c0b8053-d605-4419-81fa-e7f17c45e068","Type":"ContainerStarted","Data":"d23076c9dc83b2d98cc50b8cf1cf194f69dc27e5c78550d1fbef7cdd23135bde"} Oct 04 03:21:49 crc kubenswrapper[4742]: I1004 03:21:49.153667 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-r985t" event={"ID":"b43429d7-ac32-4ab9-91e1-1ac6010312ec","Type":"ContainerStarted","Data":"4379f428fd477a350686aad7aa1f6e937a72fb44e6d5e3b915292d820efe0ce7"} Oct 04 03:21:49 crc kubenswrapper[4742]: I1004 03:21:49.155027 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-b4g9q" event={"ID":"c32d6520-f0f0-4ba4-8317-0208036377db","Type":"ContainerStarted","Data":"5ed0e9e10f4527ae49679283d8aa7913ac2d732e03d7242790c3420858015998"} Oct 04 03:21:50 crc kubenswrapper[4742]: I1004 03:21:50.163134 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-r985t" event={"ID":"b43429d7-ac32-4ab9-91e1-1ac6010312ec","Type":"ContainerStarted","Data":"771287fcb776cdd35f5334aff545ce105dbc4f97a54212c4bba05aa9936423cd"} Oct 04 03:21:51 crc kubenswrapper[4742]: I1004 03:21:51.181253 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-cdc8r" event={"ID":"9c0b8053-d605-4419-81fa-e7f17c45e068","Type":"ContainerStarted","Data":"5902f3dd03ce9760cdcc4e5063bb523f13694bb98d5ae3f1a3f5d716628fe702"} Oct 04 03:21:51 crc kubenswrapper[4742]: I1004 03:21:51.182319 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-cdc8r" Oct 04 03:21:51 crc kubenswrapper[4742]: I1004 03:21:51.217088 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-cdc8r" podStartSLOduration=1.246660967 podStartE2EDuration="3.217062202s" podCreationTimestamp="2025-10-04 03:21:48 +0000 UTC" firstStartedPulling="2025-10-04 03:21:48.762911714 +0000 UTC m=+674.695248890" lastFinishedPulling="2025-10-04 03:21:50.733312949 +0000 UTC m=+676.665650125" observedRunningTime="2025-10-04 03:21:51.212455894 +0000 UTC m=+677.144793070" watchObservedRunningTime="2025-10-04 03:21:51.217062202 +0000 UTC m=+677.149399388" Oct 04 03:21:52 crc kubenswrapper[4742]: I1004 03:21:52.201021 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-r985t" event={"ID":"b43429d7-ac32-4ab9-91e1-1ac6010312ec","Type":"ContainerStarted","Data":"418de4f8d92286b9d3252e953e2fc60447ba12ace39dee31655dac34812f73c1"} Oct 04 03:21:52 crc kubenswrapper[4742]: I1004 03:21:52.221636 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-r985t" podStartSLOduration=2.586151284 podStartE2EDuration="4.221620003s" podCreationTimestamp="2025-10-04 03:21:48 +0000 UTC" firstStartedPulling="2025-10-04 03:21:49.2506547 +0000 UTC m=+675.182991876" lastFinishedPulling="2025-10-04 03:21:50.886123419 +0000 UTC m=+676.818460595" observedRunningTime="2025-10-04 03:21:52.219195311 +0000 UTC m=+678.151532507" watchObservedRunningTime="2025-10-04 03:21:52.221620003 +0000 UTC m=+678.153957179" Oct 04 03:21:53 crc kubenswrapper[4742]: I1004 03:21:53.217898 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-r985t" Oct 04 03:21:56 crc kubenswrapper[4742]: I1004 03:21:56.240814 4742 generic.go:334] "Generic (PLEG): container finished" podID="37f0f686-5a17-4d4a-87d8-0f1d3410530f" containerID="20471631832e8ee4a10c2a107b9b173a35612fc6f8bd40c8c02d7112a45d4b60" exitCode=0 Oct 04 03:21:56 crc kubenswrapper[4742]: I1004 03:21:56.240870 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zc8db" event={"ID":"37f0f686-5a17-4d4a-87d8-0f1d3410530f","Type":"ContainerDied","Data":"20471631832e8ee4a10c2a107b9b173a35612fc6f8bd40c8c02d7112a45d4b60"} Oct 04 03:21:56 crc kubenswrapper[4742]: I1004 03:21:56.243817 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-b4g9q" event={"ID":"c32d6520-f0f0-4ba4-8317-0208036377db","Type":"ContainerStarted","Data":"e2c19b07b5c8b0cba96e1db60894b2d3e5eecd4510dab3d11dc997e6ca573c3d"} Oct 04 03:21:56 crc kubenswrapper[4742]: I1004 03:21:56.244013 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-b4g9q" Oct 04 03:21:56 crc kubenswrapper[4742]: I1004 03:21:56.284402 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-b4g9q" podStartSLOduration=1.668462892 podStartE2EDuration="8.284381908s" podCreationTimestamp="2025-10-04 03:21:48 +0000 UTC" firstStartedPulling="2025-10-04 03:21:48.646481102 +0000 UTC m=+674.578818278" lastFinishedPulling="2025-10-04 03:21:55.262400118 +0000 UTC m=+681.194737294" observedRunningTime="2025-10-04 03:21:56.280919289 +0000 UTC m=+682.213256475" watchObservedRunningTime="2025-10-04 03:21:56.284381908 +0000 UTC m=+682.216719084" Oct 04 03:21:57 crc kubenswrapper[4742]: I1004 03:21:57.254380 4742 generic.go:334] "Generic (PLEG): container finished" podID="37f0f686-5a17-4d4a-87d8-0f1d3410530f" containerID="a0273a0627c464cdef0b00d32b87370683cd15902c90ff759bc8657ba0a07926" exitCode=0 Oct 04 03:21:57 crc kubenswrapper[4742]: I1004 03:21:57.254448 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zc8db" event={"ID":"37f0f686-5a17-4d4a-87d8-0f1d3410530f","Type":"ContainerDied","Data":"a0273a0627c464cdef0b00d32b87370683cd15902c90ff759bc8657ba0a07926"} Oct 04 03:21:58 crc kubenswrapper[4742]: I1004 03:21:58.261810 4742 generic.go:334] "Generic (PLEG): container finished" podID="37f0f686-5a17-4d4a-87d8-0f1d3410530f" containerID="a480c861ea632009aad16f4538d306b540196b525e383e93ec6f2c034b6ad15e" exitCode=0 Oct 04 03:21:58 crc kubenswrapper[4742]: I1004 03:21:58.261851 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zc8db" event={"ID":"37f0f686-5a17-4d4a-87d8-0f1d3410530f","Type":"ContainerDied","Data":"a480c861ea632009aad16f4538d306b540196b525e383e93ec6f2c034b6ad15e"} Oct 04 03:21:58 crc kubenswrapper[4742]: I1004 03:21:58.453415 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-cdc8r" Oct 04 03:21:59 crc kubenswrapper[4742]: I1004 03:21:59.065562 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-r985t" Oct 04 03:21:59 crc kubenswrapper[4742]: I1004 03:21:59.288524 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zc8db" event={"ID":"37f0f686-5a17-4d4a-87d8-0f1d3410530f","Type":"ContainerStarted","Data":"ced050c46a2a9e390d2c4554f9e586ca61f7c9d136c698a14a85ab44e9cdb8c3"} Oct 04 03:21:59 crc kubenswrapper[4742]: I1004 03:21:59.288569 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zc8db" event={"ID":"37f0f686-5a17-4d4a-87d8-0f1d3410530f","Type":"ContainerStarted","Data":"291329f3d4470e6791c38a1964143cf8015cd8cb4df672fcf87e2f6ed6a394be"} Oct 04 03:21:59 crc kubenswrapper[4742]: I1004 03:21:59.288587 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zc8db" event={"ID":"37f0f686-5a17-4d4a-87d8-0f1d3410530f","Type":"ContainerStarted","Data":"49be837adfa61bf9935b4ed0af3e52bee87525ebf7dc075d5c9354a9b2265054"} Oct 04 03:21:59 crc kubenswrapper[4742]: I1004 03:21:59.288599 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zc8db" event={"ID":"37f0f686-5a17-4d4a-87d8-0f1d3410530f","Type":"ContainerStarted","Data":"23bd51eb8e7323697a1112dc910e88c823019a795c76904f5b58d71d9bf9f699"} Oct 04 03:21:59 crc kubenswrapper[4742]: I1004 03:21:59.288610 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zc8db" event={"ID":"37f0f686-5a17-4d4a-87d8-0f1d3410530f","Type":"ContainerStarted","Data":"dce7e513f6db33198f116f1220b45a7f66c55d33fb022806a6d7bf8b23e2b740"} Oct 04 03:21:59 crc kubenswrapper[4742]: I1004 03:21:59.288620 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zc8db" event={"ID":"37f0f686-5a17-4d4a-87d8-0f1d3410530f","Type":"ContainerStarted","Data":"6f4612125d00ced02a8a007d17061af1e0d5d0a56e96becd31bed7077a9957f7"} Oct 04 03:21:59 crc kubenswrapper[4742]: I1004 03:21:59.288705 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-zc8db" Oct 04 03:21:59 crc kubenswrapper[4742]: I1004 03:21:59.317932 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-zc8db" podStartSLOduration=4.63497645 podStartE2EDuration="11.317913505s" podCreationTimestamp="2025-10-04 03:21:48 +0000 UTC" firstStartedPulling="2025-10-04 03:21:48.594580493 +0000 UTC m=+674.526917669" lastFinishedPulling="2025-10-04 03:21:55.277517548 +0000 UTC m=+681.209854724" observedRunningTime="2025-10-04 03:21:59.316983661 +0000 UTC m=+685.249320847" watchObservedRunningTime="2025-10-04 03:21:59.317913505 +0000 UTC m=+685.250250681" Oct 04 03:22:03 crc kubenswrapper[4742]: I1004 03:22:03.380894 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-zc8db" Oct 04 03:22:03 crc kubenswrapper[4742]: I1004 03:22:03.438791 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-zc8db" Oct 04 03:22:08 crc kubenswrapper[4742]: I1004 03:22:08.379577 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-b4g9q" Oct 04 03:22:08 crc kubenswrapper[4742]: I1004 03:22:08.384522 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-zc8db" Oct 04 03:22:08 crc kubenswrapper[4742]: I1004 03:22:08.442514 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-bsptq"] Oct 04 03:22:08 crc kubenswrapper[4742]: I1004 03:22:08.445051 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-bsptq" Oct 04 03:22:08 crc kubenswrapper[4742]: I1004 03:22:08.451847 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 04 03:22:08 crc kubenswrapper[4742]: I1004 03:22:08.452527 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 04 03:22:08 crc kubenswrapper[4742]: I1004 03:22:08.453571 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-bsptq"] Oct 04 03:22:08 crc kubenswrapper[4742]: I1004 03:22:08.454749 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-index-dockercfg-bptmd" Oct 04 03:22:08 crc kubenswrapper[4742]: I1004 03:22:08.594506 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2knh2\" (UniqueName: \"kubernetes.io/projected/732c29f3-87d6-4079-9837-189207dbb65a-kube-api-access-2knh2\") pod \"mariadb-operator-index-bsptq\" (UID: \"732c29f3-87d6-4079-9837-189207dbb65a\") " pod="openstack-operators/mariadb-operator-index-bsptq" Oct 04 03:22:08 crc kubenswrapper[4742]: I1004 03:22:08.696375 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2knh2\" (UniqueName: \"kubernetes.io/projected/732c29f3-87d6-4079-9837-189207dbb65a-kube-api-access-2knh2\") pod \"mariadb-operator-index-bsptq\" (UID: \"732c29f3-87d6-4079-9837-189207dbb65a\") " pod="openstack-operators/mariadb-operator-index-bsptq" Oct 04 03:22:08 crc kubenswrapper[4742]: I1004 03:22:08.723824 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2knh2\" (UniqueName: \"kubernetes.io/projected/732c29f3-87d6-4079-9837-189207dbb65a-kube-api-access-2knh2\") pod \"mariadb-operator-index-bsptq\" (UID: \"732c29f3-87d6-4079-9837-189207dbb65a\") " pod="openstack-operators/mariadb-operator-index-bsptq" Oct 04 03:22:08 crc kubenswrapper[4742]: I1004 03:22:08.772036 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-bsptq" Oct 04 03:22:09 crc kubenswrapper[4742]: I1004 03:22:09.219057 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-bsptq"] Oct 04 03:22:09 crc kubenswrapper[4742]: I1004 03:22:09.359911 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-bsptq" event={"ID":"732c29f3-87d6-4079-9837-189207dbb65a","Type":"ContainerStarted","Data":"54c11eee3670ad7c33580f98fadbe6a6065235a8924f5a2fd2620d7b433b41d0"} Oct 04 03:22:11 crc kubenswrapper[4742]: I1004 03:22:11.371859 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-bsptq" event={"ID":"732c29f3-87d6-4079-9837-189207dbb65a","Type":"ContainerStarted","Data":"f80f0827ea434fb13ff33b3fec816918e0f8d50d0c131c58f06df6aeb25ba974"} Oct 04 03:22:11 crc kubenswrapper[4742]: I1004 03:22:11.392350 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-bsptq" podStartSLOduration=2.142908219 podStartE2EDuration="3.392324704s" podCreationTimestamp="2025-10-04 03:22:08 +0000 UTC" firstStartedPulling="2025-10-04 03:22:09.235647246 +0000 UTC m=+695.167984422" lastFinishedPulling="2025-10-04 03:22:10.485063731 +0000 UTC m=+696.417400907" observedRunningTime="2025-10-04 03:22:11.387088309 +0000 UTC m=+697.319425505" watchObservedRunningTime="2025-10-04 03:22:11.392324704 +0000 UTC m=+697.324661890" Oct 04 03:22:13 crc kubenswrapper[4742]: I1004 03:22:13.620468 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-bsptq"] Oct 04 03:22:13 crc kubenswrapper[4742]: I1004 03:22:13.620653 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-index-bsptq" podUID="732c29f3-87d6-4079-9837-189207dbb65a" containerName="registry-server" containerID="cri-o://f80f0827ea434fb13ff33b3fec816918e0f8d50d0c131c58f06df6aeb25ba974" gracePeriod=2 Oct 04 03:22:14 crc kubenswrapper[4742]: I1004 03:22:14.037794 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-bsptq" Oct 04 03:22:14 crc kubenswrapper[4742]: I1004 03:22:14.178016 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2knh2\" (UniqueName: \"kubernetes.io/projected/732c29f3-87d6-4079-9837-189207dbb65a-kube-api-access-2knh2\") pod \"732c29f3-87d6-4079-9837-189207dbb65a\" (UID: \"732c29f3-87d6-4079-9837-189207dbb65a\") " Oct 04 03:22:14 crc kubenswrapper[4742]: I1004 03:22:14.183985 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/732c29f3-87d6-4079-9837-189207dbb65a-kube-api-access-2knh2" (OuterVolumeSpecName: "kube-api-access-2knh2") pod "732c29f3-87d6-4079-9837-189207dbb65a" (UID: "732c29f3-87d6-4079-9837-189207dbb65a"). InnerVolumeSpecName "kube-api-access-2knh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:22:14 crc kubenswrapper[4742]: I1004 03:22:14.225720 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-s9z66"] Oct 04 03:22:14 crc kubenswrapper[4742]: E1004 03:22:14.226032 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="732c29f3-87d6-4079-9837-189207dbb65a" containerName="registry-server" Oct 04 03:22:14 crc kubenswrapper[4742]: I1004 03:22:14.226053 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="732c29f3-87d6-4079-9837-189207dbb65a" containerName="registry-server" Oct 04 03:22:14 crc kubenswrapper[4742]: I1004 03:22:14.226216 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="732c29f3-87d6-4079-9837-189207dbb65a" containerName="registry-server" Oct 04 03:22:14 crc kubenswrapper[4742]: I1004 03:22:14.226826 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-s9z66" Oct 04 03:22:14 crc kubenswrapper[4742]: I1004 03:22:14.238715 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-s9z66"] Oct 04 03:22:14 crc kubenswrapper[4742]: I1004 03:22:14.279840 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2knh2\" (UniqueName: \"kubernetes.io/projected/732c29f3-87d6-4079-9837-189207dbb65a-kube-api-access-2knh2\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:14 crc kubenswrapper[4742]: I1004 03:22:14.381592 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7vf6\" (UniqueName: \"kubernetes.io/projected/5cc2c58e-80c3-48e8-a5f8-3fc7a8326232-kube-api-access-x7vf6\") pod \"mariadb-operator-index-s9z66\" (UID: \"5cc2c58e-80c3-48e8-a5f8-3fc7a8326232\") " pod="openstack-operators/mariadb-operator-index-s9z66" Oct 04 03:22:14 crc kubenswrapper[4742]: I1004 03:22:14.390831 4742 generic.go:334] "Generic (PLEG): container finished" podID="732c29f3-87d6-4079-9837-189207dbb65a" containerID="f80f0827ea434fb13ff33b3fec816918e0f8d50d0c131c58f06df6aeb25ba974" exitCode=0 Oct 04 03:22:14 crc kubenswrapper[4742]: I1004 03:22:14.390876 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-bsptq" event={"ID":"732c29f3-87d6-4079-9837-189207dbb65a","Type":"ContainerDied","Data":"f80f0827ea434fb13ff33b3fec816918e0f8d50d0c131c58f06df6aeb25ba974"} Oct 04 03:22:14 crc kubenswrapper[4742]: I1004 03:22:14.390899 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-bsptq" event={"ID":"732c29f3-87d6-4079-9837-189207dbb65a","Type":"ContainerDied","Data":"54c11eee3670ad7c33580f98fadbe6a6065235a8924f5a2fd2620d7b433b41d0"} Oct 04 03:22:14 crc kubenswrapper[4742]: I1004 03:22:14.390918 4742 scope.go:117] "RemoveContainer" containerID="f80f0827ea434fb13ff33b3fec816918e0f8d50d0c131c58f06df6aeb25ba974" Oct 04 03:22:14 crc kubenswrapper[4742]: I1004 03:22:14.391013 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-bsptq" Oct 04 03:22:14 crc kubenswrapper[4742]: I1004 03:22:14.410914 4742 scope.go:117] "RemoveContainer" containerID="f80f0827ea434fb13ff33b3fec816918e0f8d50d0c131c58f06df6aeb25ba974" Oct 04 03:22:14 crc kubenswrapper[4742]: E1004 03:22:14.412033 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f80f0827ea434fb13ff33b3fec816918e0f8d50d0c131c58f06df6aeb25ba974\": container with ID starting with f80f0827ea434fb13ff33b3fec816918e0f8d50d0c131c58f06df6aeb25ba974 not found: ID does not exist" containerID="f80f0827ea434fb13ff33b3fec816918e0f8d50d0c131c58f06df6aeb25ba974" Oct 04 03:22:14 crc kubenswrapper[4742]: I1004 03:22:14.412067 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f80f0827ea434fb13ff33b3fec816918e0f8d50d0c131c58f06df6aeb25ba974"} err="failed to get container status \"f80f0827ea434fb13ff33b3fec816918e0f8d50d0c131c58f06df6aeb25ba974\": rpc error: code = NotFound desc = could not find container \"f80f0827ea434fb13ff33b3fec816918e0f8d50d0c131c58f06df6aeb25ba974\": container with ID starting with f80f0827ea434fb13ff33b3fec816918e0f8d50d0c131c58f06df6aeb25ba974 not found: ID does not exist" Oct 04 03:22:14 crc kubenswrapper[4742]: I1004 03:22:14.421776 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-bsptq"] Oct 04 03:22:14 crc kubenswrapper[4742]: I1004 03:22:14.424429 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-index-bsptq"] Oct 04 03:22:14 crc kubenswrapper[4742]: I1004 03:22:14.483155 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7vf6\" (UniqueName: \"kubernetes.io/projected/5cc2c58e-80c3-48e8-a5f8-3fc7a8326232-kube-api-access-x7vf6\") pod \"mariadb-operator-index-s9z66\" (UID: \"5cc2c58e-80c3-48e8-a5f8-3fc7a8326232\") " pod="openstack-operators/mariadb-operator-index-s9z66" Oct 04 03:22:14 crc kubenswrapper[4742]: I1004 03:22:14.512422 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7vf6\" (UniqueName: \"kubernetes.io/projected/5cc2c58e-80c3-48e8-a5f8-3fc7a8326232-kube-api-access-x7vf6\") pod \"mariadb-operator-index-s9z66\" (UID: \"5cc2c58e-80c3-48e8-a5f8-3fc7a8326232\") " pod="openstack-operators/mariadb-operator-index-s9z66" Oct 04 03:22:14 crc kubenswrapper[4742]: I1004 03:22:14.557334 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-s9z66" Oct 04 03:22:14 crc kubenswrapper[4742]: I1004 03:22:14.864961 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="732c29f3-87d6-4079-9837-189207dbb65a" path="/var/lib/kubelet/pods/732c29f3-87d6-4079-9837-189207dbb65a/volumes" Oct 04 03:22:15 crc kubenswrapper[4742]: I1004 03:22:15.043780 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-s9z66"] Oct 04 03:22:15 crc kubenswrapper[4742]: W1004 03:22:15.057909 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5cc2c58e_80c3_48e8_a5f8_3fc7a8326232.slice/crio-704bcdd80a4072aa354bbca8461b55b700466424a84d1ea3eefbc425a1e5f259 WatchSource:0}: Error finding container 704bcdd80a4072aa354bbca8461b55b700466424a84d1ea3eefbc425a1e5f259: Status 404 returned error can't find the container with id 704bcdd80a4072aa354bbca8461b55b700466424a84d1ea3eefbc425a1e5f259 Oct 04 03:22:15 crc kubenswrapper[4742]: I1004 03:22:15.399965 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-s9z66" event={"ID":"5cc2c58e-80c3-48e8-a5f8-3fc7a8326232","Type":"ContainerStarted","Data":"704bcdd80a4072aa354bbca8461b55b700466424a84d1ea3eefbc425a1e5f259"} Oct 04 03:22:16 crc kubenswrapper[4742]: I1004 03:22:16.408744 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-s9z66" event={"ID":"5cc2c58e-80c3-48e8-a5f8-3fc7a8326232","Type":"ContainerStarted","Data":"babb81befd8ecb1f55c2d3de90454ce6afdc1ad3b59dfa541d792de9f94bc00e"} Oct 04 03:22:16 crc kubenswrapper[4742]: I1004 03:22:16.430394 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-s9z66" podStartSLOduration=1.9767150500000001 podStartE2EDuration="2.430362776s" podCreationTimestamp="2025-10-04 03:22:14 +0000 UTC" firstStartedPulling="2025-10-04 03:22:15.062385304 +0000 UTC m=+700.994722480" lastFinishedPulling="2025-10-04 03:22:15.51603303 +0000 UTC m=+701.448370206" observedRunningTime="2025-10-04 03:22:16.427348088 +0000 UTC m=+702.359685304" watchObservedRunningTime="2025-10-04 03:22:16.430362776 +0000 UTC m=+702.362699992" Oct 04 03:22:24 crc kubenswrapper[4742]: I1004 03:22:24.558312 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-index-s9z66" Oct 04 03:22:24 crc kubenswrapper[4742]: I1004 03:22:24.559146 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/mariadb-operator-index-s9z66" Oct 04 03:22:24 crc kubenswrapper[4742]: I1004 03:22:24.587252 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/mariadb-operator-index-s9z66" Oct 04 03:22:25 crc kubenswrapper[4742]: I1004 03:22:25.508454 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-index-s9z66" Oct 04 03:22:26 crc kubenswrapper[4742]: I1004 03:22:26.279720 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7"] Oct 04 03:22:26 crc kubenswrapper[4742]: I1004 03:22:26.281294 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7" Oct 04 03:22:26 crc kubenswrapper[4742]: I1004 03:22:26.293590 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7"] Oct 04 03:22:26 crc kubenswrapper[4742]: I1004 03:22:26.334802 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-625gw" Oct 04 03:22:26 crc kubenswrapper[4742]: I1004 03:22:26.457636 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c6aa45a1-997c-4038-97cb-12db2e9ca192-bundle\") pod \"5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7\" (UID: \"c6aa45a1-997c-4038-97cb-12db2e9ca192\") " pod="openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7" Oct 04 03:22:26 crc kubenswrapper[4742]: I1004 03:22:26.457700 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c6aa45a1-997c-4038-97cb-12db2e9ca192-util\") pod \"5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7\" (UID: \"c6aa45a1-997c-4038-97cb-12db2e9ca192\") " pod="openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7" Oct 04 03:22:26 crc kubenswrapper[4742]: I1004 03:22:26.457737 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5n2v\" (UniqueName: \"kubernetes.io/projected/c6aa45a1-997c-4038-97cb-12db2e9ca192-kube-api-access-g5n2v\") pod \"5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7\" (UID: \"c6aa45a1-997c-4038-97cb-12db2e9ca192\") " pod="openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7" Oct 04 03:22:26 crc kubenswrapper[4742]: I1004 03:22:26.559397 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c6aa45a1-997c-4038-97cb-12db2e9ca192-bundle\") pod \"5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7\" (UID: \"c6aa45a1-997c-4038-97cb-12db2e9ca192\") " pod="openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7" Oct 04 03:22:26 crc kubenswrapper[4742]: I1004 03:22:26.559497 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c6aa45a1-997c-4038-97cb-12db2e9ca192-util\") pod \"5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7\" (UID: \"c6aa45a1-997c-4038-97cb-12db2e9ca192\") " pod="openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7" Oct 04 03:22:26 crc kubenswrapper[4742]: I1004 03:22:26.559559 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5n2v\" (UniqueName: \"kubernetes.io/projected/c6aa45a1-997c-4038-97cb-12db2e9ca192-kube-api-access-g5n2v\") pod \"5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7\" (UID: \"c6aa45a1-997c-4038-97cb-12db2e9ca192\") " pod="openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7" Oct 04 03:22:26 crc kubenswrapper[4742]: I1004 03:22:26.560254 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c6aa45a1-997c-4038-97cb-12db2e9ca192-util\") pod \"5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7\" (UID: \"c6aa45a1-997c-4038-97cb-12db2e9ca192\") " pod="openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7" Oct 04 03:22:26 crc kubenswrapper[4742]: I1004 03:22:26.561077 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c6aa45a1-997c-4038-97cb-12db2e9ca192-bundle\") pod \"5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7\" (UID: \"c6aa45a1-997c-4038-97cb-12db2e9ca192\") " pod="openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7" Oct 04 03:22:26 crc kubenswrapper[4742]: I1004 03:22:26.596020 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5n2v\" (UniqueName: \"kubernetes.io/projected/c6aa45a1-997c-4038-97cb-12db2e9ca192-kube-api-access-g5n2v\") pod \"5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7\" (UID: \"c6aa45a1-997c-4038-97cb-12db2e9ca192\") " pod="openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7" Oct 04 03:22:26 crc kubenswrapper[4742]: I1004 03:22:26.649320 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7" Oct 04 03:22:27 crc kubenswrapper[4742]: I1004 03:22:27.077712 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7"] Oct 04 03:22:27 crc kubenswrapper[4742]: I1004 03:22:27.489914 4742 generic.go:334] "Generic (PLEG): container finished" podID="c6aa45a1-997c-4038-97cb-12db2e9ca192" containerID="628ceed9e5a30dff06546ecd8ec5f471051c7509f5f6cc6e9d5384a9e998705e" exitCode=0 Oct 04 03:22:27 crc kubenswrapper[4742]: I1004 03:22:27.489965 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7" event={"ID":"c6aa45a1-997c-4038-97cb-12db2e9ca192","Type":"ContainerDied","Data":"628ceed9e5a30dff06546ecd8ec5f471051c7509f5f6cc6e9d5384a9e998705e"} Oct 04 03:22:27 crc kubenswrapper[4742]: I1004 03:22:27.489995 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7" event={"ID":"c6aa45a1-997c-4038-97cb-12db2e9ca192","Type":"ContainerStarted","Data":"423c0d218e7be48f80d0011a22b4d5084e31e818c80d8d6dccf1e5baf0f2affc"} Oct 04 03:22:28 crc kubenswrapper[4742]: I1004 03:22:28.495692 4742 generic.go:334] "Generic (PLEG): container finished" podID="c6aa45a1-997c-4038-97cb-12db2e9ca192" containerID="02be5bbe23d19aec5f9630b77960197bc6f4d28baf9cd5ff305a5f4b21e331d7" exitCode=0 Oct 04 03:22:28 crc kubenswrapper[4742]: I1004 03:22:28.495849 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7" event={"ID":"c6aa45a1-997c-4038-97cb-12db2e9ca192","Type":"ContainerDied","Data":"02be5bbe23d19aec5f9630b77960197bc6f4d28baf9cd5ff305a5f4b21e331d7"} Oct 04 03:22:29 crc kubenswrapper[4742]: I1004 03:22:29.503143 4742 generic.go:334] "Generic (PLEG): container finished" podID="c6aa45a1-997c-4038-97cb-12db2e9ca192" containerID="d5d3c7c079670d34426df73e801c885df4639d59ec73399c9a516a252eddf2c9" exitCode=0 Oct 04 03:22:29 crc kubenswrapper[4742]: I1004 03:22:29.503195 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7" event={"ID":"c6aa45a1-997c-4038-97cb-12db2e9ca192","Type":"ContainerDied","Data":"d5d3c7c079670d34426df73e801c885df4639d59ec73399c9a516a252eddf2c9"} Oct 04 03:22:30 crc kubenswrapper[4742]: I1004 03:22:30.785757 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7" Oct 04 03:22:30 crc kubenswrapper[4742]: I1004 03:22:30.924679 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c6aa45a1-997c-4038-97cb-12db2e9ca192-bundle\") pod \"c6aa45a1-997c-4038-97cb-12db2e9ca192\" (UID: \"c6aa45a1-997c-4038-97cb-12db2e9ca192\") " Oct 04 03:22:30 crc kubenswrapper[4742]: I1004 03:22:30.924803 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c6aa45a1-997c-4038-97cb-12db2e9ca192-util\") pod \"c6aa45a1-997c-4038-97cb-12db2e9ca192\" (UID: \"c6aa45a1-997c-4038-97cb-12db2e9ca192\") " Oct 04 03:22:30 crc kubenswrapper[4742]: I1004 03:22:30.924928 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5n2v\" (UniqueName: \"kubernetes.io/projected/c6aa45a1-997c-4038-97cb-12db2e9ca192-kube-api-access-g5n2v\") pod \"c6aa45a1-997c-4038-97cb-12db2e9ca192\" (UID: \"c6aa45a1-997c-4038-97cb-12db2e9ca192\") " Oct 04 03:22:30 crc kubenswrapper[4742]: I1004 03:22:30.925696 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6aa45a1-997c-4038-97cb-12db2e9ca192-bundle" (OuterVolumeSpecName: "bundle") pod "c6aa45a1-997c-4038-97cb-12db2e9ca192" (UID: "c6aa45a1-997c-4038-97cb-12db2e9ca192"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:22:30 crc kubenswrapper[4742]: I1004 03:22:30.932907 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6aa45a1-997c-4038-97cb-12db2e9ca192-kube-api-access-g5n2v" (OuterVolumeSpecName: "kube-api-access-g5n2v") pod "c6aa45a1-997c-4038-97cb-12db2e9ca192" (UID: "c6aa45a1-997c-4038-97cb-12db2e9ca192"). InnerVolumeSpecName "kube-api-access-g5n2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:22:30 crc kubenswrapper[4742]: I1004 03:22:30.942627 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6aa45a1-997c-4038-97cb-12db2e9ca192-util" (OuterVolumeSpecName: "util") pod "c6aa45a1-997c-4038-97cb-12db2e9ca192" (UID: "c6aa45a1-997c-4038-97cb-12db2e9ca192"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:22:31 crc kubenswrapper[4742]: I1004 03:22:31.028339 4742 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c6aa45a1-997c-4038-97cb-12db2e9ca192-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:31 crc kubenswrapper[4742]: I1004 03:22:31.028737 4742 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c6aa45a1-997c-4038-97cb-12db2e9ca192-util\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:31 crc kubenswrapper[4742]: I1004 03:22:31.029144 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5n2v\" (UniqueName: \"kubernetes.io/projected/c6aa45a1-997c-4038-97cb-12db2e9ca192-kube-api-access-g5n2v\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:31 crc kubenswrapper[4742]: I1004 03:22:31.525911 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7" event={"ID":"c6aa45a1-997c-4038-97cb-12db2e9ca192","Type":"ContainerDied","Data":"423c0d218e7be48f80d0011a22b4d5084e31e818c80d8d6dccf1e5baf0f2affc"} Oct 04 03:22:31 crc kubenswrapper[4742]: I1004 03:22:31.526015 4742 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="423c0d218e7be48f80d0011a22b4d5084e31e818c80d8d6dccf1e5baf0f2affc" Oct 04 03:22:31 crc kubenswrapper[4742]: I1004 03:22:31.526115 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7" Oct 04 03:22:36 crc kubenswrapper[4742]: I1004 03:22:36.038804 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-86bc99bbf9-j65wm"] Oct 04 03:22:36 crc kubenswrapper[4742]: E1004 03:22:36.040146 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6aa45a1-997c-4038-97cb-12db2e9ca192" containerName="extract" Oct 04 03:22:36 crc kubenswrapper[4742]: I1004 03:22:36.040168 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6aa45a1-997c-4038-97cb-12db2e9ca192" containerName="extract" Oct 04 03:22:36 crc kubenswrapper[4742]: E1004 03:22:36.040196 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6aa45a1-997c-4038-97cb-12db2e9ca192" containerName="util" Oct 04 03:22:36 crc kubenswrapper[4742]: I1004 03:22:36.040205 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6aa45a1-997c-4038-97cb-12db2e9ca192" containerName="util" Oct 04 03:22:36 crc kubenswrapper[4742]: E1004 03:22:36.040233 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6aa45a1-997c-4038-97cb-12db2e9ca192" containerName="pull" Oct 04 03:22:36 crc kubenswrapper[4742]: I1004 03:22:36.040243 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6aa45a1-997c-4038-97cb-12db2e9ca192" containerName="pull" Oct 04 03:22:36 crc kubenswrapper[4742]: I1004 03:22:36.040405 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6aa45a1-997c-4038-97cb-12db2e9ca192" containerName="extract" Oct 04 03:22:36 crc kubenswrapper[4742]: I1004 03:22:36.041373 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-86bc99bbf9-j65wm" Oct 04 03:22:36 crc kubenswrapper[4742]: I1004 03:22:36.042901 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 04 03:22:36 crc kubenswrapper[4742]: I1004 03:22:36.044526 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-service-cert" Oct 04 03:22:36 crc kubenswrapper[4742]: I1004 03:22:36.044731 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-t9cdx" Oct 04 03:22:36 crc kubenswrapper[4742]: I1004 03:22:36.056209 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-86bc99bbf9-j65wm"] Oct 04 03:22:36 crc kubenswrapper[4742]: I1004 03:22:36.131793 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcnlg\" (UniqueName: \"kubernetes.io/projected/e298bed2-0909-45f9-b422-21e118fc389a-kube-api-access-gcnlg\") pod \"mariadb-operator-controller-manager-86bc99bbf9-j65wm\" (UID: \"e298bed2-0909-45f9-b422-21e118fc389a\") " pod="openstack-operators/mariadb-operator-controller-manager-86bc99bbf9-j65wm" Oct 04 03:22:36 crc kubenswrapper[4742]: I1004 03:22:36.131988 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e298bed2-0909-45f9-b422-21e118fc389a-webhook-cert\") pod \"mariadb-operator-controller-manager-86bc99bbf9-j65wm\" (UID: \"e298bed2-0909-45f9-b422-21e118fc389a\") " pod="openstack-operators/mariadb-operator-controller-manager-86bc99bbf9-j65wm" Oct 04 03:22:36 crc kubenswrapper[4742]: I1004 03:22:36.132127 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e298bed2-0909-45f9-b422-21e118fc389a-apiservice-cert\") pod \"mariadb-operator-controller-manager-86bc99bbf9-j65wm\" (UID: \"e298bed2-0909-45f9-b422-21e118fc389a\") " pod="openstack-operators/mariadb-operator-controller-manager-86bc99bbf9-j65wm" Oct 04 03:22:36 crc kubenswrapper[4742]: I1004 03:22:36.234403 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e298bed2-0909-45f9-b422-21e118fc389a-webhook-cert\") pod \"mariadb-operator-controller-manager-86bc99bbf9-j65wm\" (UID: \"e298bed2-0909-45f9-b422-21e118fc389a\") " pod="openstack-operators/mariadb-operator-controller-manager-86bc99bbf9-j65wm" Oct 04 03:22:36 crc kubenswrapper[4742]: I1004 03:22:36.234499 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e298bed2-0909-45f9-b422-21e118fc389a-apiservice-cert\") pod \"mariadb-operator-controller-manager-86bc99bbf9-j65wm\" (UID: \"e298bed2-0909-45f9-b422-21e118fc389a\") " pod="openstack-operators/mariadb-operator-controller-manager-86bc99bbf9-j65wm" Oct 04 03:22:36 crc kubenswrapper[4742]: I1004 03:22:36.234617 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcnlg\" (UniqueName: \"kubernetes.io/projected/e298bed2-0909-45f9-b422-21e118fc389a-kube-api-access-gcnlg\") pod \"mariadb-operator-controller-manager-86bc99bbf9-j65wm\" (UID: \"e298bed2-0909-45f9-b422-21e118fc389a\") " pod="openstack-operators/mariadb-operator-controller-manager-86bc99bbf9-j65wm" Oct 04 03:22:36 crc kubenswrapper[4742]: I1004 03:22:36.252344 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e298bed2-0909-45f9-b422-21e118fc389a-webhook-cert\") pod \"mariadb-operator-controller-manager-86bc99bbf9-j65wm\" (UID: \"e298bed2-0909-45f9-b422-21e118fc389a\") " pod="openstack-operators/mariadb-operator-controller-manager-86bc99bbf9-j65wm" Oct 04 03:22:36 crc kubenswrapper[4742]: I1004 03:22:36.253109 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e298bed2-0909-45f9-b422-21e118fc389a-apiservice-cert\") pod \"mariadb-operator-controller-manager-86bc99bbf9-j65wm\" (UID: \"e298bed2-0909-45f9-b422-21e118fc389a\") " pod="openstack-operators/mariadb-operator-controller-manager-86bc99bbf9-j65wm" Oct 04 03:22:36 crc kubenswrapper[4742]: I1004 03:22:36.258956 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcnlg\" (UniqueName: \"kubernetes.io/projected/e298bed2-0909-45f9-b422-21e118fc389a-kube-api-access-gcnlg\") pod \"mariadb-operator-controller-manager-86bc99bbf9-j65wm\" (UID: \"e298bed2-0909-45f9-b422-21e118fc389a\") " pod="openstack-operators/mariadb-operator-controller-manager-86bc99bbf9-j65wm" Oct 04 03:22:36 crc kubenswrapper[4742]: I1004 03:22:36.358593 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-86bc99bbf9-j65wm" Oct 04 03:22:36 crc kubenswrapper[4742]: I1004 03:22:36.645616 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-86bc99bbf9-j65wm"] Oct 04 03:22:36 crc kubenswrapper[4742]: W1004 03:22:36.661916 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode298bed2_0909_45f9_b422_21e118fc389a.slice/crio-3d6cd2a748c6a62e3e676334cdc6f9fba996cd23a811a8a0d258293ed0110c63 WatchSource:0}: Error finding container 3d6cd2a748c6a62e3e676334cdc6f9fba996cd23a811a8a0d258293ed0110c63: Status 404 returned error can't find the container with id 3d6cd2a748c6a62e3e676334cdc6f9fba996cd23a811a8a0d258293ed0110c63 Oct 04 03:22:37 crc kubenswrapper[4742]: I1004 03:22:37.573078 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-86bc99bbf9-j65wm" event={"ID":"e298bed2-0909-45f9-b422-21e118fc389a","Type":"ContainerStarted","Data":"3d6cd2a748c6a62e3e676334cdc6f9fba996cd23a811a8a0d258293ed0110c63"} Oct 04 03:22:41 crc kubenswrapper[4742]: I1004 03:22:41.615516 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-86bc99bbf9-j65wm" event={"ID":"e298bed2-0909-45f9-b422-21e118fc389a","Type":"ContainerStarted","Data":"b729fa9efa41dc45f48be567e03adb7e1ca559fb8bd1517128ec6ed0b1ebbbc5"} Oct 04 03:22:43 crc kubenswrapper[4742]: I1004 03:22:43.628583 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-86bc99bbf9-j65wm" event={"ID":"e298bed2-0909-45f9-b422-21e118fc389a","Type":"ContainerStarted","Data":"985b47f3e647586f9b768edaa8016418d51f0b8d4f86c84b98fa3865ae7e1911"} Oct 04 03:22:43 crc kubenswrapper[4742]: I1004 03:22:43.629080 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-86bc99bbf9-j65wm" Oct 04 03:22:43 crc kubenswrapper[4742]: I1004 03:22:43.647938 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-86bc99bbf9-j65wm" podStartSLOduration=1.348820923 podStartE2EDuration="7.647897659s" podCreationTimestamp="2025-10-04 03:22:36 +0000 UTC" firstStartedPulling="2025-10-04 03:22:36.665795891 +0000 UTC m=+722.598133067" lastFinishedPulling="2025-10-04 03:22:42.964872617 +0000 UTC m=+728.897209803" observedRunningTime="2025-10-04 03:22:43.646541434 +0000 UTC m=+729.578878610" watchObservedRunningTime="2025-10-04 03:22:43.647897659 +0000 UTC m=+729.580234835" Oct 04 03:22:46 crc kubenswrapper[4742]: I1004 03:22:46.365482 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-86bc99bbf9-j65wm" Oct 04 03:22:47 crc kubenswrapper[4742]: I1004 03:22:47.793107 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-xt57c"] Oct 04 03:22:47 crc kubenswrapper[4742]: I1004 03:22:47.793927 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-xt57c" Oct 04 03:22:47 crc kubenswrapper[4742]: I1004 03:22:47.800004 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-index-dockercfg-v2ttg" Oct 04 03:22:47 crc kubenswrapper[4742]: I1004 03:22:47.852963 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-xt57c"] Oct 04 03:22:47 crc kubenswrapper[4742]: I1004 03:22:47.933891 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ns9g8\" (UniqueName: \"kubernetes.io/projected/48d9b79b-aefd-4990-a54e-66f9dadb7b02-kube-api-access-ns9g8\") pod \"infra-operator-index-xt57c\" (UID: \"48d9b79b-aefd-4990-a54e-66f9dadb7b02\") " pod="openstack-operators/infra-operator-index-xt57c" Oct 04 03:22:48 crc kubenswrapper[4742]: I1004 03:22:48.035518 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ns9g8\" (UniqueName: \"kubernetes.io/projected/48d9b79b-aefd-4990-a54e-66f9dadb7b02-kube-api-access-ns9g8\") pod \"infra-operator-index-xt57c\" (UID: \"48d9b79b-aefd-4990-a54e-66f9dadb7b02\") " pod="openstack-operators/infra-operator-index-xt57c" Oct 04 03:22:48 crc kubenswrapper[4742]: I1004 03:22:48.067544 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ns9g8\" (UniqueName: \"kubernetes.io/projected/48d9b79b-aefd-4990-a54e-66f9dadb7b02-kube-api-access-ns9g8\") pod \"infra-operator-index-xt57c\" (UID: \"48d9b79b-aefd-4990-a54e-66f9dadb7b02\") " pod="openstack-operators/infra-operator-index-xt57c" Oct 04 03:22:48 crc kubenswrapper[4742]: I1004 03:22:48.110756 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-xt57c" Oct 04 03:22:48 crc kubenswrapper[4742]: I1004 03:22:48.674452 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-xt57c"] Oct 04 03:22:49 crc kubenswrapper[4742]: I1004 03:22:49.674083 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-xt57c" event={"ID":"48d9b79b-aefd-4990-a54e-66f9dadb7b02","Type":"ContainerStarted","Data":"9b2b4b551a7287edc081b9121d835cff2cf2b8c77a92c752ed6493566e65b2a6"} Oct 04 03:22:50 crc kubenswrapper[4742]: I1004 03:22:50.226795 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-xt57c"] Oct 04 03:22:50 crc kubenswrapper[4742]: I1004 03:22:50.631619 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-rlwxl"] Oct 04 03:22:50 crc kubenswrapper[4742]: I1004 03:22:50.633300 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-rlwxl" Oct 04 03:22:50 crc kubenswrapper[4742]: I1004 03:22:50.641541 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-rlwxl"] Oct 04 03:22:50 crc kubenswrapper[4742]: I1004 03:22:50.688814 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-xt57c" event={"ID":"48d9b79b-aefd-4990-a54e-66f9dadb7b02","Type":"ContainerStarted","Data":"6d1acbd8c6970e0a094edeb6adb9ef93a6dbc549310e102453e73bb47451cd6d"} Oct 04 03:22:50 crc kubenswrapper[4742]: I1004 03:22:50.710244 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-xt57c" podStartSLOduration=2.70667367 podStartE2EDuration="3.710221355s" podCreationTimestamp="2025-10-04 03:22:47 +0000 UTC" firstStartedPulling="2025-10-04 03:22:48.693263 +0000 UTC m=+734.625600176" lastFinishedPulling="2025-10-04 03:22:49.696810685 +0000 UTC m=+735.629147861" observedRunningTime="2025-10-04 03:22:50.709758573 +0000 UTC m=+736.642095759" watchObservedRunningTime="2025-10-04 03:22:50.710221355 +0000 UTC m=+736.642558541" Oct 04 03:22:50 crc kubenswrapper[4742]: I1004 03:22:50.782807 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpsnv\" (UniqueName: \"kubernetes.io/projected/1ba2be5e-e2e5-45db-a77e-cbad59fee6a3-kube-api-access-kpsnv\") pod \"infra-operator-index-rlwxl\" (UID: \"1ba2be5e-e2e5-45db-a77e-cbad59fee6a3\") " pod="openstack-operators/infra-operator-index-rlwxl" Oct 04 03:22:50 crc kubenswrapper[4742]: I1004 03:22:50.884200 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpsnv\" (UniqueName: \"kubernetes.io/projected/1ba2be5e-e2e5-45db-a77e-cbad59fee6a3-kube-api-access-kpsnv\") pod \"infra-operator-index-rlwxl\" (UID: \"1ba2be5e-e2e5-45db-a77e-cbad59fee6a3\") " pod="openstack-operators/infra-operator-index-rlwxl" Oct 04 03:22:50 crc kubenswrapper[4742]: I1004 03:22:50.914391 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpsnv\" (UniqueName: \"kubernetes.io/projected/1ba2be5e-e2e5-45db-a77e-cbad59fee6a3-kube-api-access-kpsnv\") pod \"infra-operator-index-rlwxl\" (UID: \"1ba2be5e-e2e5-45db-a77e-cbad59fee6a3\") " pod="openstack-operators/infra-operator-index-rlwxl" Oct 04 03:22:50 crc kubenswrapper[4742]: I1004 03:22:50.998700 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-rlwxl" Oct 04 03:22:51 crc kubenswrapper[4742]: I1004 03:22:51.508242 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-rlwxl"] Oct 04 03:22:51 crc kubenswrapper[4742]: W1004 03:22:51.519743 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ba2be5e_e2e5_45db_a77e_cbad59fee6a3.slice/crio-a1e603e95b205c5c9584e131e5385d3da267cd8bf5682dc294b29674d3f48d57 WatchSource:0}: Error finding container a1e603e95b205c5c9584e131e5385d3da267cd8bf5682dc294b29674d3f48d57: Status 404 returned error can't find the container with id a1e603e95b205c5c9584e131e5385d3da267cd8bf5682dc294b29674d3f48d57 Oct 04 03:22:51 crc kubenswrapper[4742]: I1004 03:22:51.696572 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-index-xt57c" podUID="48d9b79b-aefd-4990-a54e-66f9dadb7b02" containerName="registry-server" containerID="cri-o://6d1acbd8c6970e0a094edeb6adb9ef93a6dbc549310e102453e73bb47451cd6d" gracePeriod=2 Oct 04 03:22:51 crc kubenswrapper[4742]: I1004 03:22:51.696907 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-rlwxl" event={"ID":"1ba2be5e-e2e5-45db-a77e-cbad59fee6a3","Type":"ContainerStarted","Data":"a1e603e95b205c5c9584e131e5385d3da267cd8bf5682dc294b29674d3f48d57"} Oct 04 03:22:52 crc kubenswrapper[4742]: I1004 03:22:52.159062 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-xt57c" Oct 04 03:22:52 crc kubenswrapper[4742]: I1004 03:22:52.317108 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ns9g8\" (UniqueName: \"kubernetes.io/projected/48d9b79b-aefd-4990-a54e-66f9dadb7b02-kube-api-access-ns9g8\") pod \"48d9b79b-aefd-4990-a54e-66f9dadb7b02\" (UID: \"48d9b79b-aefd-4990-a54e-66f9dadb7b02\") " Oct 04 03:22:52 crc kubenswrapper[4742]: I1004 03:22:52.328594 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48d9b79b-aefd-4990-a54e-66f9dadb7b02-kube-api-access-ns9g8" (OuterVolumeSpecName: "kube-api-access-ns9g8") pod "48d9b79b-aefd-4990-a54e-66f9dadb7b02" (UID: "48d9b79b-aefd-4990-a54e-66f9dadb7b02"). InnerVolumeSpecName "kube-api-access-ns9g8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:22:52 crc kubenswrapper[4742]: I1004 03:22:52.418993 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ns9g8\" (UniqueName: \"kubernetes.io/projected/48d9b79b-aefd-4990-a54e-66f9dadb7b02-kube-api-access-ns9g8\") on node \"crc\" DevicePath \"\"" Oct 04 03:22:52 crc kubenswrapper[4742]: I1004 03:22:52.707110 4742 generic.go:334] "Generic (PLEG): container finished" podID="48d9b79b-aefd-4990-a54e-66f9dadb7b02" containerID="6d1acbd8c6970e0a094edeb6adb9ef93a6dbc549310e102453e73bb47451cd6d" exitCode=0 Oct 04 03:22:52 crc kubenswrapper[4742]: I1004 03:22:52.707205 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-xt57c" event={"ID":"48d9b79b-aefd-4990-a54e-66f9dadb7b02","Type":"ContainerDied","Data":"6d1acbd8c6970e0a094edeb6adb9ef93a6dbc549310e102453e73bb47451cd6d"} Oct 04 03:22:52 crc kubenswrapper[4742]: I1004 03:22:52.708981 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-xt57c" event={"ID":"48d9b79b-aefd-4990-a54e-66f9dadb7b02","Type":"ContainerDied","Data":"9b2b4b551a7287edc081b9121d835cff2cf2b8c77a92c752ed6493566e65b2a6"} Oct 04 03:22:52 crc kubenswrapper[4742]: I1004 03:22:52.709033 4742 scope.go:117] "RemoveContainer" containerID="6d1acbd8c6970e0a094edeb6adb9ef93a6dbc549310e102453e73bb47451cd6d" Oct 04 03:22:52 crc kubenswrapper[4742]: I1004 03:22:52.707245 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-xt57c" Oct 04 03:22:52 crc kubenswrapper[4742]: I1004 03:22:52.721288 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-rlwxl" event={"ID":"1ba2be5e-e2e5-45db-a77e-cbad59fee6a3","Type":"ContainerStarted","Data":"7643a0875ae0856828cf2c7250d76fb88696915c9cc2c76cef36d1bc283052fd"} Oct 04 03:22:52 crc kubenswrapper[4742]: I1004 03:22:52.755250 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-rlwxl" podStartSLOduration=2.323710058 podStartE2EDuration="2.755214354s" podCreationTimestamp="2025-10-04 03:22:50 +0000 UTC" firstStartedPulling="2025-10-04 03:22:51.525512837 +0000 UTC m=+737.457850023" lastFinishedPulling="2025-10-04 03:22:51.957017123 +0000 UTC m=+737.889354319" observedRunningTime="2025-10-04 03:22:52.750488002 +0000 UTC m=+738.682825178" watchObservedRunningTime="2025-10-04 03:22:52.755214354 +0000 UTC m=+738.687551570" Oct 04 03:22:52 crc kubenswrapper[4742]: I1004 03:22:52.759793 4742 scope.go:117] "RemoveContainer" containerID="6d1acbd8c6970e0a094edeb6adb9ef93a6dbc549310e102453e73bb47451cd6d" Oct 04 03:22:52 crc kubenswrapper[4742]: E1004 03:22:52.760574 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d1acbd8c6970e0a094edeb6adb9ef93a6dbc549310e102453e73bb47451cd6d\": container with ID starting with 6d1acbd8c6970e0a094edeb6adb9ef93a6dbc549310e102453e73bb47451cd6d not found: ID does not exist" containerID="6d1acbd8c6970e0a094edeb6adb9ef93a6dbc549310e102453e73bb47451cd6d" Oct 04 03:22:52 crc kubenswrapper[4742]: I1004 03:22:52.760646 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d1acbd8c6970e0a094edeb6adb9ef93a6dbc549310e102453e73bb47451cd6d"} err="failed to get container status \"6d1acbd8c6970e0a094edeb6adb9ef93a6dbc549310e102453e73bb47451cd6d\": rpc error: code = NotFound desc = could not find container \"6d1acbd8c6970e0a094edeb6adb9ef93a6dbc549310e102453e73bb47451cd6d\": container with ID starting with 6d1acbd8c6970e0a094edeb6adb9ef93a6dbc549310e102453e73bb47451cd6d not found: ID does not exist" Oct 04 03:22:52 crc kubenswrapper[4742]: I1004 03:22:52.770755 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-xt57c"] Oct 04 03:22:52 crc kubenswrapper[4742]: I1004 03:22:52.775731 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/infra-operator-index-xt57c"] Oct 04 03:22:52 crc kubenswrapper[4742]: I1004 03:22:52.866554 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48d9b79b-aefd-4990-a54e-66f9dadb7b02" path="/var/lib/kubelet/pods/48d9b79b-aefd-4990-a54e-66f9dadb7b02/volumes" Oct 04 03:23:01 crc kubenswrapper[4742]: I1004 03:23:00.998997 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-index-rlwxl" Oct 04 03:23:01 crc kubenswrapper[4742]: I1004 03:23:01.000081 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/infra-operator-index-rlwxl" Oct 04 03:23:01 crc kubenswrapper[4742]: I1004 03:23:01.041836 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/infra-operator-index-rlwxl" Oct 04 03:23:01 crc kubenswrapper[4742]: I1004 03:23:01.835894 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-index-rlwxl" Oct 04 03:23:02 crc kubenswrapper[4742]: I1004 03:23:02.757648 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-gszdm"] Oct 04 03:23:02 crc kubenswrapper[4742]: I1004 03:23:02.758594 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-gszdm" podUID="690f72b6-3754-442a-b1c9-59ff8ed83c1a" containerName="controller-manager" containerID="cri-o://d0849799409f769f9737527b6af9087a31e3db241ce43e7c56a8db8797e9a877" gracePeriod=30 Oct 04 03:23:02 crc kubenswrapper[4742]: I1004 03:23:02.781404 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-t7pnt"] Oct 04 03:23:02 crc kubenswrapper[4742]: I1004 03:23:02.781718 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t7pnt" podUID="938dc4e6-1064-4eab-a3e3-8e3096109655" containerName="route-controller-manager" containerID="cri-o://61b050e50e078ef054514389b9665c66a71facfe81ec7126e3dde6a61bd818e7" gracePeriod=30 Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.319858 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-gszdm" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.372533 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t7pnt" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.391734 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/690f72b6-3754-442a-b1c9-59ff8ed83c1a-serving-cert\") pod \"690f72b6-3754-442a-b1c9-59ff8ed83c1a\" (UID: \"690f72b6-3754-442a-b1c9-59ff8ed83c1a\") " Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.391850 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/690f72b6-3754-442a-b1c9-59ff8ed83c1a-config\") pod \"690f72b6-3754-442a-b1c9-59ff8ed83c1a\" (UID: \"690f72b6-3754-442a-b1c9-59ff8ed83c1a\") " Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.391891 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/690f72b6-3754-442a-b1c9-59ff8ed83c1a-proxy-ca-bundles\") pod \"690f72b6-3754-442a-b1c9-59ff8ed83c1a\" (UID: \"690f72b6-3754-442a-b1c9-59ff8ed83c1a\") " Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.391927 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/690f72b6-3754-442a-b1c9-59ff8ed83c1a-client-ca\") pod \"690f72b6-3754-442a-b1c9-59ff8ed83c1a\" (UID: \"690f72b6-3754-442a-b1c9-59ff8ed83c1a\") " Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.392014 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8flcs\" (UniqueName: \"kubernetes.io/projected/690f72b6-3754-442a-b1c9-59ff8ed83c1a-kube-api-access-8flcs\") pod \"690f72b6-3754-442a-b1c9-59ff8ed83c1a\" (UID: \"690f72b6-3754-442a-b1c9-59ff8ed83c1a\") " Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.393133 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/690f72b6-3754-442a-b1c9-59ff8ed83c1a-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "690f72b6-3754-442a-b1c9-59ff8ed83c1a" (UID: "690f72b6-3754-442a-b1c9-59ff8ed83c1a"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.393195 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/690f72b6-3754-442a-b1c9-59ff8ed83c1a-client-ca" (OuterVolumeSpecName: "client-ca") pod "690f72b6-3754-442a-b1c9-59ff8ed83c1a" (UID: "690f72b6-3754-442a-b1c9-59ff8ed83c1a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.393654 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/690f72b6-3754-442a-b1c9-59ff8ed83c1a-config" (OuterVolumeSpecName: "config") pod "690f72b6-3754-442a-b1c9-59ff8ed83c1a" (UID: "690f72b6-3754-442a-b1c9-59ff8ed83c1a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.403825 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/690f72b6-3754-442a-b1c9-59ff8ed83c1a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "690f72b6-3754-442a-b1c9-59ff8ed83c1a" (UID: "690f72b6-3754-442a-b1c9-59ff8ed83c1a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.404803 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/690f72b6-3754-442a-b1c9-59ff8ed83c1a-kube-api-access-8flcs" (OuterVolumeSpecName: "kube-api-access-8flcs") pod "690f72b6-3754-442a-b1c9-59ff8ed83c1a" (UID: "690f72b6-3754-442a-b1c9-59ff8ed83c1a"). InnerVolumeSpecName "kube-api-access-8flcs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.493934 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngndk\" (UniqueName: \"kubernetes.io/projected/938dc4e6-1064-4eab-a3e3-8e3096109655-kube-api-access-ngndk\") pod \"938dc4e6-1064-4eab-a3e3-8e3096109655\" (UID: \"938dc4e6-1064-4eab-a3e3-8e3096109655\") " Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.494050 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/938dc4e6-1064-4eab-a3e3-8e3096109655-config\") pod \"938dc4e6-1064-4eab-a3e3-8e3096109655\" (UID: \"938dc4e6-1064-4eab-a3e3-8e3096109655\") " Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.494095 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/938dc4e6-1064-4eab-a3e3-8e3096109655-serving-cert\") pod \"938dc4e6-1064-4eab-a3e3-8e3096109655\" (UID: \"938dc4e6-1064-4eab-a3e3-8e3096109655\") " Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.494123 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/938dc4e6-1064-4eab-a3e3-8e3096109655-client-ca\") pod \"938dc4e6-1064-4eab-a3e3-8e3096109655\" (UID: \"938dc4e6-1064-4eab-a3e3-8e3096109655\") " Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.494447 4742 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/690f72b6-3754-442a-b1c9-59ff8ed83c1a-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.494465 4742 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/690f72b6-3754-442a-b1c9-59ff8ed83c1a-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.494476 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8flcs\" (UniqueName: \"kubernetes.io/projected/690f72b6-3754-442a-b1c9-59ff8ed83c1a-kube-api-access-8flcs\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.494487 4742 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/690f72b6-3754-442a-b1c9-59ff8ed83c1a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.494498 4742 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/690f72b6-3754-442a-b1c9-59ff8ed83c1a-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.495792 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr"] Oct 04 03:23:03 crc kubenswrapper[4742]: E1004 03:23:03.496138 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="938dc4e6-1064-4eab-a3e3-8e3096109655" containerName="route-controller-manager" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.496166 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="938dc4e6-1064-4eab-a3e3-8e3096109655" containerName="route-controller-manager" Oct 04 03:23:03 crc kubenswrapper[4742]: E1004 03:23:03.496185 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="690f72b6-3754-442a-b1c9-59ff8ed83c1a" containerName="controller-manager" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.496195 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="690f72b6-3754-442a-b1c9-59ff8ed83c1a" containerName="controller-manager" Oct 04 03:23:03 crc kubenswrapper[4742]: E1004 03:23:03.496208 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48d9b79b-aefd-4990-a54e-66f9dadb7b02" containerName="registry-server" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.496219 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="48d9b79b-aefd-4990-a54e-66f9dadb7b02" containerName="registry-server" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.496366 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="48d9b79b-aefd-4990-a54e-66f9dadb7b02" containerName="registry-server" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.496388 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="938dc4e6-1064-4eab-a3e3-8e3096109655" containerName="route-controller-manager" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.496400 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="690f72b6-3754-442a-b1c9-59ff8ed83c1a" containerName="controller-manager" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.497383 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.498568 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/938dc4e6-1064-4eab-a3e3-8e3096109655-client-ca" (OuterVolumeSpecName: "client-ca") pod "938dc4e6-1064-4eab-a3e3-8e3096109655" (UID: "938dc4e6-1064-4eab-a3e3-8e3096109655"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.498660 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/938dc4e6-1064-4eab-a3e3-8e3096109655-config" (OuterVolumeSpecName: "config") pod "938dc4e6-1064-4eab-a3e3-8e3096109655" (UID: "938dc4e6-1064-4eab-a3e3-8e3096109655"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.503005 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/938dc4e6-1064-4eab-a3e3-8e3096109655-kube-api-access-ngndk" (OuterVolumeSpecName: "kube-api-access-ngndk") pod "938dc4e6-1064-4eab-a3e3-8e3096109655" (UID: "938dc4e6-1064-4eab-a3e3-8e3096109655"). InnerVolumeSpecName "kube-api-access-ngndk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.503362 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-625gw" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.504606 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/938dc4e6-1064-4eab-a3e3-8e3096109655-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "938dc4e6-1064-4eab-a3e3-8e3096109655" (UID: "938dc4e6-1064-4eab-a3e3-8e3096109655"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.518173 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr"] Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.595756 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksvvc\" (UniqueName: \"kubernetes.io/projected/d72194dd-00e3-474f-a278-35ec718e1fd6-kube-api-access-ksvvc\") pod \"ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr\" (UID: \"d72194dd-00e3-474f-a278-35ec718e1fd6\") " pod="openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.595842 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d72194dd-00e3-474f-a278-35ec718e1fd6-bundle\") pod \"ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr\" (UID: \"d72194dd-00e3-474f-a278-35ec718e1fd6\") " pod="openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.595872 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d72194dd-00e3-474f-a278-35ec718e1fd6-util\") pod \"ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr\" (UID: \"d72194dd-00e3-474f-a278-35ec718e1fd6\") " pod="openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.595951 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngndk\" (UniqueName: \"kubernetes.io/projected/938dc4e6-1064-4eab-a3e3-8e3096109655-kube-api-access-ngndk\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.595967 4742 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/938dc4e6-1064-4eab-a3e3-8e3096109655-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.595978 4742 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/938dc4e6-1064-4eab-a3e3-8e3096109655-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.595986 4742 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/938dc4e6-1064-4eab-a3e3-8e3096109655-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.697042 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksvvc\" (UniqueName: \"kubernetes.io/projected/d72194dd-00e3-474f-a278-35ec718e1fd6-kube-api-access-ksvvc\") pod \"ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr\" (UID: \"d72194dd-00e3-474f-a278-35ec718e1fd6\") " pod="openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.697160 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d72194dd-00e3-474f-a278-35ec718e1fd6-bundle\") pod \"ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr\" (UID: \"d72194dd-00e3-474f-a278-35ec718e1fd6\") " pod="openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.697206 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d72194dd-00e3-474f-a278-35ec718e1fd6-util\") pod \"ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr\" (UID: \"d72194dd-00e3-474f-a278-35ec718e1fd6\") " pod="openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.697923 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d72194dd-00e3-474f-a278-35ec718e1fd6-util\") pod \"ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr\" (UID: \"d72194dd-00e3-474f-a278-35ec718e1fd6\") " pod="openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.698095 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d72194dd-00e3-474f-a278-35ec718e1fd6-bundle\") pod \"ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr\" (UID: \"d72194dd-00e3-474f-a278-35ec718e1fd6\") " pod="openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.722973 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksvvc\" (UniqueName: \"kubernetes.io/projected/d72194dd-00e3-474f-a278-35ec718e1fd6-kube-api-access-ksvvc\") pod \"ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr\" (UID: \"d72194dd-00e3-474f-a278-35ec718e1fd6\") " pod="openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.820599 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.825158 4742 generic.go:334] "Generic (PLEG): container finished" podID="690f72b6-3754-442a-b1c9-59ff8ed83c1a" containerID="d0849799409f769f9737527b6af9087a31e3db241ce43e7c56a8db8797e9a877" exitCode=0 Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.825320 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-gszdm" event={"ID":"690f72b6-3754-442a-b1c9-59ff8ed83c1a","Type":"ContainerDied","Data":"d0849799409f769f9737527b6af9087a31e3db241ce43e7c56a8db8797e9a877"} Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.825385 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-gszdm" event={"ID":"690f72b6-3754-442a-b1c9-59ff8ed83c1a","Type":"ContainerDied","Data":"b3c480a48a7dcbccc27a89b8e256fb821b91bfe0860bd5386db6f0644c17291d"} Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.825421 4742 scope.go:117] "RemoveContainer" containerID="d0849799409f769f9737527b6af9087a31e3db241ce43e7c56a8db8797e9a877" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.825636 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-gszdm" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.830265 4742 generic.go:334] "Generic (PLEG): container finished" podID="938dc4e6-1064-4eab-a3e3-8e3096109655" containerID="61b050e50e078ef054514389b9665c66a71facfe81ec7126e3dde6a61bd818e7" exitCode=0 Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.830345 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t7pnt" event={"ID":"938dc4e6-1064-4eab-a3e3-8e3096109655","Type":"ContainerDied","Data":"61b050e50e078ef054514389b9665c66a71facfe81ec7126e3dde6a61bd818e7"} Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.830475 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t7pnt" event={"ID":"938dc4e6-1064-4eab-a3e3-8e3096109655","Type":"ContainerDied","Data":"c103178b8682daab0e696cffcb94dc93f414347f16a5057bfe2164e47bb29ff1"} Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.830391 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t7pnt" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.878823 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-t7pnt"] Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.879298 4742 scope.go:117] "RemoveContainer" containerID="d0849799409f769f9737527b6af9087a31e3db241ce43e7c56a8db8797e9a877" Oct 04 03:23:03 crc kubenswrapper[4742]: E1004 03:23:03.879886 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0849799409f769f9737527b6af9087a31e3db241ce43e7c56a8db8797e9a877\": container with ID starting with d0849799409f769f9737527b6af9087a31e3db241ce43e7c56a8db8797e9a877 not found: ID does not exist" containerID="d0849799409f769f9737527b6af9087a31e3db241ce43e7c56a8db8797e9a877" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.879939 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0849799409f769f9737527b6af9087a31e3db241ce43e7c56a8db8797e9a877"} err="failed to get container status \"d0849799409f769f9737527b6af9087a31e3db241ce43e7c56a8db8797e9a877\": rpc error: code = NotFound desc = could not find container \"d0849799409f769f9737527b6af9087a31e3db241ce43e7c56a8db8797e9a877\": container with ID starting with d0849799409f769f9737527b6af9087a31e3db241ce43e7c56a8db8797e9a877 not found: ID does not exist" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.879979 4742 scope.go:117] "RemoveContainer" containerID="61b050e50e078ef054514389b9665c66a71facfe81ec7126e3dde6a61bd818e7" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.891980 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-t7pnt"] Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.913243 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-gszdm"] Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.919641 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-gszdm"] Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.928647 4742 scope.go:117] "RemoveContainer" containerID="61b050e50e078ef054514389b9665c66a71facfe81ec7126e3dde6a61bd818e7" Oct 04 03:23:03 crc kubenswrapper[4742]: E1004 03:23:03.929401 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61b050e50e078ef054514389b9665c66a71facfe81ec7126e3dde6a61bd818e7\": container with ID starting with 61b050e50e078ef054514389b9665c66a71facfe81ec7126e3dde6a61bd818e7 not found: ID does not exist" containerID="61b050e50e078ef054514389b9665c66a71facfe81ec7126e3dde6a61bd818e7" Oct 04 03:23:03 crc kubenswrapper[4742]: I1004 03:23:03.929464 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61b050e50e078ef054514389b9665c66a71facfe81ec7126e3dde6a61bd818e7"} err="failed to get container status \"61b050e50e078ef054514389b9665c66a71facfe81ec7126e3dde6a61bd818e7\": rpc error: code = NotFound desc = could not find container \"61b050e50e078ef054514389b9665c66a71facfe81ec7126e3dde6a61bd818e7\": container with ID starting with 61b050e50e078ef054514389b9665c66a71facfe81ec7126e3dde6a61bd818e7 not found: ID does not exist" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.106062 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr"] Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.567944 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7c569ccb78-8mgw8"] Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.569395 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c569ccb78-8mgw8" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.571099 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-64df44bff8-488hk"] Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.571947 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-64df44bff8-488hk" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.572413 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.572518 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.572818 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.575670 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.576026 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.576236 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.576371 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.576433 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.576516 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.576675 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.576783 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.578387 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.589003 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-64df44bff8-488hk"] Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.596049 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.604071 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7c569ccb78-8mgw8"] Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.720176 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fae2bb90-5081-4abd-a3e7-5a51e01e3cea-serving-cert\") pod \"controller-manager-7c569ccb78-8mgw8\" (UID: \"fae2bb90-5081-4abd-a3e7-5a51e01e3cea\") " pod="openshift-controller-manager/controller-manager-7c569ccb78-8mgw8" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.720641 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f827e4c9-4a6c-4f39-872d-854ee7de0f74-config\") pod \"route-controller-manager-64df44bff8-488hk\" (UID: \"f827e4c9-4a6c-4f39-872d-854ee7de0f74\") " pod="openshift-route-controller-manager/route-controller-manager-64df44bff8-488hk" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.720759 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vx7f\" (UniqueName: \"kubernetes.io/projected/f827e4c9-4a6c-4f39-872d-854ee7de0f74-kube-api-access-7vx7f\") pod \"route-controller-manager-64df44bff8-488hk\" (UID: \"f827e4c9-4a6c-4f39-872d-854ee7de0f74\") " pod="openshift-route-controller-manager/route-controller-manager-64df44bff8-488hk" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.720930 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fae2bb90-5081-4abd-a3e7-5a51e01e3cea-config\") pod \"controller-manager-7c569ccb78-8mgw8\" (UID: \"fae2bb90-5081-4abd-a3e7-5a51e01e3cea\") " pod="openshift-controller-manager/controller-manager-7c569ccb78-8mgw8" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.720988 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fae2bb90-5081-4abd-a3e7-5a51e01e3cea-client-ca\") pod \"controller-manager-7c569ccb78-8mgw8\" (UID: \"fae2bb90-5081-4abd-a3e7-5a51e01e3cea\") " pod="openshift-controller-manager/controller-manager-7c569ccb78-8mgw8" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.721062 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fae2bb90-5081-4abd-a3e7-5a51e01e3cea-proxy-ca-bundles\") pod \"controller-manager-7c569ccb78-8mgw8\" (UID: \"fae2bb90-5081-4abd-a3e7-5a51e01e3cea\") " pod="openshift-controller-manager/controller-manager-7c569ccb78-8mgw8" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.721135 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f827e4c9-4a6c-4f39-872d-854ee7de0f74-client-ca\") pod \"route-controller-manager-64df44bff8-488hk\" (UID: \"f827e4c9-4a6c-4f39-872d-854ee7de0f74\") " pod="openshift-route-controller-manager/route-controller-manager-64df44bff8-488hk" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.721179 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f827e4c9-4a6c-4f39-872d-854ee7de0f74-serving-cert\") pod \"route-controller-manager-64df44bff8-488hk\" (UID: \"f827e4c9-4a6c-4f39-872d-854ee7de0f74\") " pod="openshift-route-controller-manager/route-controller-manager-64df44bff8-488hk" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.721258 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thttj\" (UniqueName: \"kubernetes.io/projected/fae2bb90-5081-4abd-a3e7-5a51e01e3cea-kube-api-access-thttj\") pod \"controller-manager-7c569ccb78-8mgw8\" (UID: \"fae2bb90-5081-4abd-a3e7-5a51e01e3cea\") " pod="openshift-controller-manager/controller-manager-7c569ccb78-8mgw8" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.822867 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fae2bb90-5081-4abd-a3e7-5a51e01e3cea-serving-cert\") pod \"controller-manager-7c569ccb78-8mgw8\" (UID: \"fae2bb90-5081-4abd-a3e7-5a51e01e3cea\") " pod="openshift-controller-manager/controller-manager-7c569ccb78-8mgw8" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.823031 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f827e4c9-4a6c-4f39-872d-854ee7de0f74-config\") pod \"route-controller-manager-64df44bff8-488hk\" (UID: \"f827e4c9-4a6c-4f39-872d-854ee7de0f74\") " pod="openshift-route-controller-manager/route-controller-manager-64df44bff8-488hk" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.823219 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vx7f\" (UniqueName: \"kubernetes.io/projected/f827e4c9-4a6c-4f39-872d-854ee7de0f74-kube-api-access-7vx7f\") pod \"route-controller-manager-64df44bff8-488hk\" (UID: \"f827e4c9-4a6c-4f39-872d-854ee7de0f74\") " pod="openshift-route-controller-manager/route-controller-manager-64df44bff8-488hk" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.823337 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fae2bb90-5081-4abd-a3e7-5a51e01e3cea-config\") pod \"controller-manager-7c569ccb78-8mgw8\" (UID: \"fae2bb90-5081-4abd-a3e7-5a51e01e3cea\") " pod="openshift-controller-manager/controller-manager-7c569ccb78-8mgw8" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.823506 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fae2bb90-5081-4abd-a3e7-5a51e01e3cea-client-ca\") pod \"controller-manager-7c569ccb78-8mgw8\" (UID: \"fae2bb90-5081-4abd-a3e7-5a51e01e3cea\") " pod="openshift-controller-manager/controller-manager-7c569ccb78-8mgw8" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.823559 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fae2bb90-5081-4abd-a3e7-5a51e01e3cea-proxy-ca-bundles\") pod \"controller-manager-7c569ccb78-8mgw8\" (UID: \"fae2bb90-5081-4abd-a3e7-5a51e01e3cea\") " pod="openshift-controller-manager/controller-manager-7c569ccb78-8mgw8" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.823630 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f827e4c9-4a6c-4f39-872d-854ee7de0f74-client-ca\") pod \"route-controller-manager-64df44bff8-488hk\" (UID: \"f827e4c9-4a6c-4f39-872d-854ee7de0f74\") " pod="openshift-route-controller-manager/route-controller-manager-64df44bff8-488hk" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.823712 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f827e4c9-4a6c-4f39-872d-854ee7de0f74-serving-cert\") pod \"route-controller-manager-64df44bff8-488hk\" (UID: \"f827e4c9-4a6c-4f39-872d-854ee7de0f74\") " pod="openshift-route-controller-manager/route-controller-manager-64df44bff8-488hk" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.823838 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thttj\" (UniqueName: \"kubernetes.io/projected/fae2bb90-5081-4abd-a3e7-5a51e01e3cea-kube-api-access-thttj\") pod \"controller-manager-7c569ccb78-8mgw8\" (UID: \"fae2bb90-5081-4abd-a3e7-5a51e01e3cea\") " pod="openshift-controller-manager/controller-manager-7c569ccb78-8mgw8" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.829185 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f827e4c9-4a6c-4f39-872d-854ee7de0f74-client-ca\") pod \"route-controller-manager-64df44bff8-488hk\" (UID: \"f827e4c9-4a6c-4f39-872d-854ee7de0f74\") " pod="openshift-route-controller-manager/route-controller-manager-64df44bff8-488hk" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.830598 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fae2bb90-5081-4abd-a3e7-5a51e01e3cea-config\") pod \"controller-manager-7c569ccb78-8mgw8\" (UID: \"fae2bb90-5081-4abd-a3e7-5a51e01e3cea\") " pod="openshift-controller-manager/controller-manager-7c569ccb78-8mgw8" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.831018 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f827e4c9-4a6c-4f39-872d-854ee7de0f74-config\") pod \"route-controller-manager-64df44bff8-488hk\" (UID: \"f827e4c9-4a6c-4f39-872d-854ee7de0f74\") " pod="openshift-route-controller-manager/route-controller-manager-64df44bff8-488hk" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.831354 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fae2bb90-5081-4abd-a3e7-5a51e01e3cea-proxy-ca-bundles\") pod \"controller-manager-7c569ccb78-8mgw8\" (UID: \"fae2bb90-5081-4abd-a3e7-5a51e01e3cea\") " pod="openshift-controller-manager/controller-manager-7c569ccb78-8mgw8" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.836524 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f827e4c9-4a6c-4f39-872d-854ee7de0f74-serving-cert\") pod \"route-controller-manager-64df44bff8-488hk\" (UID: \"f827e4c9-4a6c-4f39-872d-854ee7de0f74\") " pod="openshift-route-controller-manager/route-controller-manager-64df44bff8-488hk" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.833982 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fae2bb90-5081-4abd-a3e7-5a51e01e3cea-client-ca\") pod \"controller-manager-7c569ccb78-8mgw8\" (UID: \"fae2bb90-5081-4abd-a3e7-5a51e01e3cea\") " pod="openshift-controller-manager/controller-manager-7c569ccb78-8mgw8" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.841800 4742 generic.go:334] "Generic (PLEG): container finished" podID="d72194dd-00e3-474f-a278-35ec718e1fd6" containerID="c1744bbe41b56c7ff72f1d9458a3f9ab488a673948061254a6075508e315e4b9" exitCode=0 Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.841973 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr" event={"ID":"d72194dd-00e3-474f-a278-35ec718e1fd6","Type":"ContainerDied","Data":"c1744bbe41b56c7ff72f1d9458a3f9ab488a673948061254a6075508e315e4b9"} Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.842081 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr" event={"ID":"d72194dd-00e3-474f-a278-35ec718e1fd6","Type":"ContainerStarted","Data":"a08f1782a24ec5e22c2de3873ab39643818d95427170c42a353bb02a74feb168"} Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.863201 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fae2bb90-5081-4abd-a3e7-5a51e01e3cea-serving-cert\") pod \"controller-manager-7c569ccb78-8mgw8\" (UID: \"fae2bb90-5081-4abd-a3e7-5a51e01e3cea\") " pod="openshift-controller-manager/controller-manager-7c569ccb78-8mgw8" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.868518 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thttj\" (UniqueName: \"kubernetes.io/projected/fae2bb90-5081-4abd-a3e7-5a51e01e3cea-kube-api-access-thttj\") pod \"controller-manager-7c569ccb78-8mgw8\" (UID: \"fae2bb90-5081-4abd-a3e7-5a51e01e3cea\") " pod="openshift-controller-manager/controller-manager-7c569ccb78-8mgw8" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.877843 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vx7f\" (UniqueName: \"kubernetes.io/projected/f827e4c9-4a6c-4f39-872d-854ee7de0f74-kube-api-access-7vx7f\") pod \"route-controller-manager-64df44bff8-488hk\" (UID: \"f827e4c9-4a6c-4f39-872d-854ee7de0f74\") " pod="openshift-route-controller-manager/route-controller-manager-64df44bff8-488hk" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.881807 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="690f72b6-3754-442a-b1c9-59ff8ed83c1a" path="/var/lib/kubelet/pods/690f72b6-3754-442a-b1c9-59ff8ed83c1a/volumes" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.883956 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="938dc4e6-1064-4eab-a3e3-8e3096109655" path="/var/lib/kubelet/pods/938dc4e6-1064-4eab-a3e3-8e3096109655/volumes" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.909043 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c569ccb78-8mgw8" Oct 04 03:23:04 crc kubenswrapper[4742]: I1004 03:23:04.926817 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-64df44bff8-488hk" Oct 04 03:23:05 crc kubenswrapper[4742]: I1004 03:23:05.197604 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-64df44bff8-488hk"] Oct 04 03:23:05 crc kubenswrapper[4742]: I1004 03:23:05.270350 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7c569ccb78-8mgw8"] Oct 04 03:23:05 crc kubenswrapper[4742]: W1004 03:23:05.281964 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfae2bb90_5081_4abd_a3e7_5a51e01e3cea.slice/crio-d325300c6abed8bfd4ec9897023d29789744e00c03990b5ee76829158f55d378 WatchSource:0}: Error finding container d325300c6abed8bfd4ec9897023d29789744e00c03990b5ee76829158f55d378: Status 404 returned error can't find the container with id d325300c6abed8bfd4ec9897023d29789744e00c03990b5ee76829158f55d378 Oct 04 03:23:05 crc kubenswrapper[4742]: I1004 03:23:05.874132 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-64df44bff8-488hk" event={"ID":"f827e4c9-4a6c-4f39-872d-854ee7de0f74","Type":"ContainerStarted","Data":"0a797335c4593d05d078b41c4cbe24ce98ec9707f87c938db1dbdb6b70aeeaf7"} Oct 04 03:23:05 crc kubenswrapper[4742]: I1004 03:23:05.874595 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-64df44bff8-488hk" event={"ID":"f827e4c9-4a6c-4f39-872d-854ee7de0f74","Type":"ContainerStarted","Data":"345f9e5ec06ed768545133b0fde7bf78edf505c3b90f13d19ee1cb97d3316b63"} Oct 04 03:23:05 crc kubenswrapper[4742]: I1004 03:23:05.875296 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-64df44bff8-488hk" Oct 04 03:23:05 crc kubenswrapper[4742]: I1004 03:23:05.876757 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c569ccb78-8mgw8" event={"ID":"fae2bb90-5081-4abd-a3e7-5a51e01e3cea","Type":"ContainerStarted","Data":"a340cdb35991d5be77cf07079639ce7d6059b27366e4c9be33a858e2dbf9a72c"} Oct 04 03:23:05 crc kubenswrapper[4742]: I1004 03:23:05.876793 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c569ccb78-8mgw8" event={"ID":"fae2bb90-5081-4abd-a3e7-5a51e01e3cea","Type":"ContainerStarted","Data":"d325300c6abed8bfd4ec9897023d29789744e00c03990b5ee76829158f55d378"} Oct 04 03:23:05 crc kubenswrapper[4742]: I1004 03:23:05.877256 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7c569ccb78-8mgw8" Oct 04 03:23:05 crc kubenswrapper[4742]: I1004 03:23:05.879245 4742 generic.go:334] "Generic (PLEG): container finished" podID="d72194dd-00e3-474f-a278-35ec718e1fd6" containerID="c5caea6e662b419cfded7b13bc7074ef8c22b1ceb56bf0d20db55e6f25b0ef2e" exitCode=0 Oct 04 03:23:05 crc kubenswrapper[4742]: I1004 03:23:05.879291 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr" event={"ID":"d72194dd-00e3-474f-a278-35ec718e1fd6","Type":"ContainerDied","Data":"c5caea6e662b419cfded7b13bc7074ef8c22b1ceb56bf0d20db55e6f25b0ef2e"} Oct 04 03:23:05 crc kubenswrapper[4742]: I1004 03:23:05.881490 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-64df44bff8-488hk" Oct 04 03:23:05 crc kubenswrapper[4742]: I1004 03:23:05.886220 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7c569ccb78-8mgw8" Oct 04 03:23:05 crc kubenswrapper[4742]: I1004 03:23:05.917747 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-64df44bff8-488hk" podStartSLOduration=2.91772594 podStartE2EDuration="2.91772594s" podCreationTimestamp="2025-10-04 03:23:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:23:05.898944756 +0000 UTC m=+751.831281922" watchObservedRunningTime="2025-10-04 03:23:05.91772594 +0000 UTC m=+751.850063106" Oct 04 03:23:05 crc kubenswrapper[4742]: I1004 03:23:05.919093 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7c569ccb78-8mgw8" podStartSLOduration=2.919088065 podStartE2EDuration="2.919088065s" podCreationTimestamp="2025-10-04 03:23:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:23:05.916642232 +0000 UTC m=+751.848979398" watchObservedRunningTime="2025-10-04 03:23:05.919088065 +0000 UTC m=+751.851425241" Oct 04 03:23:06 crc kubenswrapper[4742]: I1004 03:23:06.891034 4742 generic.go:334] "Generic (PLEG): container finished" podID="d72194dd-00e3-474f-a278-35ec718e1fd6" containerID="d1ab18276f3cc25da7020dae86584de58d12577d2cdb280fb41d95cc1a801f07" exitCode=0 Oct 04 03:23:06 crc kubenswrapper[4742]: I1004 03:23:06.891438 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr" event={"ID":"d72194dd-00e3-474f-a278-35ec718e1fd6","Type":"ContainerDied","Data":"d1ab18276f3cc25da7020dae86584de58d12577d2cdb280fb41d95cc1a801f07"} Oct 04 03:23:08 crc kubenswrapper[4742]: I1004 03:23:08.296882 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr" Oct 04 03:23:08 crc kubenswrapper[4742]: I1004 03:23:08.385187 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksvvc\" (UniqueName: \"kubernetes.io/projected/d72194dd-00e3-474f-a278-35ec718e1fd6-kube-api-access-ksvvc\") pod \"d72194dd-00e3-474f-a278-35ec718e1fd6\" (UID: \"d72194dd-00e3-474f-a278-35ec718e1fd6\") " Oct 04 03:23:08 crc kubenswrapper[4742]: I1004 03:23:08.385386 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d72194dd-00e3-474f-a278-35ec718e1fd6-util\") pod \"d72194dd-00e3-474f-a278-35ec718e1fd6\" (UID: \"d72194dd-00e3-474f-a278-35ec718e1fd6\") " Oct 04 03:23:08 crc kubenswrapper[4742]: I1004 03:23:08.385462 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d72194dd-00e3-474f-a278-35ec718e1fd6-bundle\") pod \"d72194dd-00e3-474f-a278-35ec718e1fd6\" (UID: \"d72194dd-00e3-474f-a278-35ec718e1fd6\") " Oct 04 03:23:08 crc kubenswrapper[4742]: I1004 03:23:08.386726 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d72194dd-00e3-474f-a278-35ec718e1fd6-bundle" (OuterVolumeSpecName: "bundle") pod "d72194dd-00e3-474f-a278-35ec718e1fd6" (UID: "d72194dd-00e3-474f-a278-35ec718e1fd6"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:23:08 crc kubenswrapper[4742]: I1004 03:23:08.396030 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d72194dd-00e3-474f-a278-35ec718e1fd6-kube-api-access-ksvvc" (OuterVolumeSpecName: "kube-api-access-ksvvc") pod "d72194dd-00e3-474f-a278-35ec718e1fd6" (UID: "d72194dd-00e3-474f-a278-35ec718e1fd6"). InnerVolumeSpecName "kube-api-access-ksvvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:23:08 crc kubenswrapper[4742]: I1004 03:23:08.412798 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d72194dd-00e3-474f-a278-35ec718e1fd6-util" (OuterVolumeSpecName: "util") pod "d72194dd-00e3-474f-a278-35ec718e1fd6" (UID: "d72194dd-00e3-474f-a278-35ec718e1fd6"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:23:08 crc kubenswrapper[4742]: I1004 03:23:08.488089 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ksvvc\" (UniqueName: \"kubernetes.io/projected/d72194dd-00e3-474f-a278-35ec718e1fd6-kube-api-access-ksvvc\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:08 crc kubenswrapper[4742]: I1004 03:23:08.488142 4742 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d72194dd-00e3-474f-a278-35ec718e1fd6-util\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:08 crc kubenswrapper[4742]: I1004 03:23:08.488156 4742 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d72194dd-00e3-474f-a278-35ec718e1fd6-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:08 crc kubenswrapper[4742]: I1004 03:23:08.909106 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr" event={"ID":"d72194dd-00e3-474f-a278-35ec718e1fd6","Type":"ContainerDied","Data":"a08f1782a24ec5e22c2de3873ab39643818d95427170c42a353bb02a74feb168"} Oct 04 03:23:08 crc kubenswrapper[4742]: I1004 03:23:08.909693 4742 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a08f1782a24ec5e22c2de3873ab39643818d95427170c42a353bb02a74feb168" Oct 04 03:23:08 crc kubenswrapper[4742]: I1004 03:23:08.909170 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr" Oct 04 03:23:09 crc kubenswrapper[4742]: I1004 03:23:09.066401 4742 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.381204 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Oct 04 03:23:12 crc kubenswrapper[4742]: E1004 03:23:12.382100 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d72194dd-00e3-474f-a278-35ec718e1fd6" containerName="pull" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.382123 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="d72194dd-00e3-474f-a278-35ec718e1fd6" containerName="pull" Oct 04 03:23:12 crc kubenswrapper[4742]: E1004 03:23:12.382147 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d72194dd-00e3-474f-a278-35ec718e1fd6" containerName="extract" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.382157 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="d72194dd-00e3-474f-a278-35ec718e1fd6" containerName="extract" Oct 04 03:23:12 crc kubenswrapper[4742]: E1004 03:23:12.382184 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d72194dd-00e3-474f-a278-35ec718e1fd6" containerName="util" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.382194 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="d72194dd-00e3-474f-a278-35ec718e1fd6" containerName="util" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.382414 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="d72194dd-00e3-474f-a278-35ec718e1fd6" containerName="extract" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.383464 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.387984 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openshift-service-ca.crt" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.388065 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"galera-openstack-dockercfg-dt6pl" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.388477 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"osp-secret" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.392687 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.392929 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config-data" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.393196 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"kube-root-ca.crt" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.398532 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.400137 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.404309 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.416132 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.417911 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.436971 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.450925 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.453646 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1b69f4d7-789b-4992-9b33-a839e3105797-config-data-default\") pod \"openstack-galera-0\" (UID: \"1b69f4d7-789b-4992-9b33-a839e3105797\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.453700 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1b69f4d7-789b-4992-9b33-a839e3105797-operator-scripts\") pod \"openstack-galera-0\" (UID: \"1b69f4d7-789b-4992-9b33-a839e3105797\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.453722 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/68095fb7-c650-4e91-aa84-1a66a90076db-kolla-config\") pod \"openstack-galera-1\" (UID: \"68095fb7-c650-4e91-aa84-1a66a90076db\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.453751 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/68095fb7-c650-4e91-aa84-1a66a90076db-config-data-generated\") pod \"openstack-galera-1\" (UID: \"68095fb7-c650-4e91-aa84-1a66a90076db\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.453771 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h54dz\" (UniqueName: \"kubernetes.io/projected/68095fb7-c650-4e91-aa84-1a66a90076db-kube-api-access-h54dz\") pod \"openstack-galera-1\" (UID: \"68095fb7-c650-4e91-aa84-1a66a90076db\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.453789 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cz4s\" (UniqueName: \"kubernetes.io/projected/1b69f4d7-789b-4992-9b33-a839e3105797-kube-api-access-5cz4s\") pod \"openstack-galera-0\" (UID: \"1b69f4d7-789b-4992-9b33-a839e3105797\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.453811 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68095fb7-c650-4e91-aa84-1a66a90076db-operator-scripts\") pod \"openstack-galera-1\" (UID: \"68095fb7-c650-4e91-aa84-1a66a90076db\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.453841 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1b69f4d7-789b-4992-9b33-a839e3105797-kolla-config\") pod \"openstack-galera-0\" (UID: \"1b69f4d7-789b-4992-9b33-a839e3105797\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.453861 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"68095fb7-c650-4e91-aa84-1a66a90076db\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.453879 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/68095fb7-c650-4e91-aa84-1a66a90076db-secrets\") pod \"openstack-galera-1\" (UID: \"68095fb7-c650-4e91-aa84-1a66a90076db\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.453898 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"1b69f4d7-789b-4992-9b33-a839e3105797\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.453928 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/1b69f4d7-789b-4992-9b33-a839e3105797-secrets\") pod \"openstack-galera-0\" (UID: \"1b69f4d7-789b-4992-9b33-a839e3105797\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.453947 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1b69f4d7-789b-4992-9b33-a839e3105797-config-data-generated\") pod \"openstack-galera-0\" (UID: \"1b69f4d7-789b-4992-9b33-a839e3105797\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.453971 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/68095fb7-c650-4e91-aa84-1a66a90076db-config-data-default\") pod \"openstack-galera-1\" (UID: \"68095fb7-c650-4e91-aa84-1a66a90076db\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.555379 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/68095fb7-c650-4e91-aa84-1a66a90076db-config-data-generated\") pod \"openstack-galera-1\" (UID: \"68095fb7-c650-4e91-aa84-1a66a90076db\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.555450 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cz4s\" (UniqueName: \"kubernetes.io/projected/1b69f4d7-789b-4992-9b33-a839e3105797-kube-api-access-5cz4s\") pod \"openstack-galera-0\" (UID: \"1b69f4d7-789b-4992-9b33-a839e3105797\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.555477 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h54dz\" (UniqueName: \"kubernetes.io/projected/68095fb7-c650-4e91-aa84-1a66a90076db-kube-api-access-h54dz\") pod \"openstack-galera-1\" (UID: \"68095fb7-c650-4e91-aa84-1a66a90076db\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.555510 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4786b0c4-6846-4873-a3fb-0e155d15e2d0-config-data-default\") pod \"openstack-galera-2\" (UID: \"4786b0c4-6846-4873-a3fb-0e155d15e2d0\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.555543 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-2\" (UID: \"4786b0c4-6846-4873-a3fb-0e155d15e2d0\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.555573 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68095fb7-c650-4e91-aa84-1a66a90076db-operator-scripts\") pod \"openstack-galera-1\" (UID: \"68095fb7-c650-4e91-aa84-1a66a90076db\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.555608 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4786b0c4-6846-4873-a3fb-0e155d15e2d0-operator-scripts\") pod \"openstack-galera-2\" (UID: \"4786b0c4-6846-4873-a3fb-0e155d15e2d0\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.555636 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1b69f4d7-789b-4992-9b33-a839e3105797-kolla-config\") pod \"openstack-galera-0\" (UID: \"1b69f4d7-789b-4992-9b33-a839e3105797\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.555792 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"68095fb7-c650-4e91-aa84-1a66a90076db\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.555888 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/68095fb7-c650-4e91-aa84-1a66a90076db-secrets\") pod \"openstack-galera-1\" (UID: \"68095fb7-c650-4e91-aa84-1a66a90076db\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.555948 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"1b69f4d7-789b-4992-9b33-a839e3105797\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.555947 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/68095fb7-c650-4e91-aa84-1a66a90076db-config-data-generated\") pod \"openstack-galera-1\" (UID: \"68095fb7-c650-4e91-aa84-1a66a90076db\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.556243 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4786b0c4-6846-4873-a3fb-0e155d15e2d0-config-data-generated\") pod \"openstack-galera-2\" (UID: \"4786b0c4-6846-4873-a3fb-0e155d15e2d0\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.556321 4742 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"1b69f4d7-789b-4992-9b33-a839e3105797\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/openstack-galera-0" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.556331 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4786b0c4-6846-4873-a3fb-0e155d15e2d0-kolla-config\") pod \"openstack-galera-2\" (UID: \"4786b0c4-6846-4873-a3fb-0e155d15e2d0\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.556321 4742 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"68095fb7-c650-4e91-aa84-1a66a90076db\") device mount path \"/mnt/openstack/pv03\"" pod="glance-kuttl-tests/openstack-galera-1" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.556670 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/1b69f4d7-789b-4992-9b33-a839e3105797-secrets\") pod \"openstack-galera-0\" (UID: \"1b69f4d7-789b-4992-9b33-a839e3105797\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.556717 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1b69f4d7-789b-4992-9b33-a839e3105797-config-data-generated\") pod \"openstack-galera-0\" (UID: \"1b69f4d7-789b-4992-9b33-a839e3105797\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.556763 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/4786b0c4-6846-4873-a3fb-0e155d15e2d0-secrets\") pod \"openstack-galera-2\" (UID: \"4786b0c4-6846-4873-a3fb-0e155d15e2d0\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.556802 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glpv4\" (UniqueName: \"kubernetes.io/projected/4786b0c4-6846-4873-a3fb-0e155d15e2d0-kube-api-access-glpv4\") pod \"openstack-galera-2\" (UID: \"4786b0c4-6846-4873-a3fb-0e155d15e2d0\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.556832 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/68095fb7-c650-4e91-aa84-1a66a90076db-config-data-default\") pod \"openstack-galera-1\" (UID: \"68095fb7-c650-4e91-aa84-1a66a90076db\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.556856 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1b69f4d7-789b-4992-9b33-a839e3105797-config-data-default\") pod \"openstack-galera-0\" (UID: \"1b69f4d7-789b-4992-9b33-a839e3105797\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.556873 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1b69f4d7-789b-4992-9b33-a839e3105797-kolla-config\") pod \"openstack-galera-0\" (UID: \"1b69f4d7-789b-4992-9b33-a839e3105797\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.556894 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1b69f4d7-789b-4992-9b33-a839e3105797-operator-scripts\") pod \"openstack-galera-0\" (UID: \"1b69f4d7-789b-4992-9b33-a839e3105797\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.556923 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/68095fb7-c650-4e91-aa84-1a66a90076db-kolla-config\") pod \"openstack-galera-1\" (UID: \"68095fb7-c650-4e91-aa84-1a66a90076db\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.557343 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1b69f4d7-789b-4992-9b33-a839e3105797-config-data-generated\") pod \"openstack-galera-0\" (UID: \"1b69f4d7-789b-4992-9b33-a839e3105797\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.557578 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/68095fb7-c650-4e91-aa84-1a66a90076db-kolla-config\") pod \"openstack-galera-1\" (UID: \"68095fb7-c650-4e91-aa84-1a66a90076db\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.557822 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68095fb7-c650-4e91-aa84-1a66a90076db-operator-scripts\") pod \"openstack-galera-1\" (UID: \"68095fb7-c650-4e91-aa84-1a66a90076db\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.558249 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1b69f4d7-789b-4992-9b33-a839e3105797-config-data-default\") pod \"openstack-galera-0\" (UID: \"1b69f4d7-789b-4992-9b33-a839e3105797\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.558365 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/68095fb7-c650-4e91-aa84-1a66a90076db-config-data-default\") pod \"openstack-galera-1\" (UID: \"68095fb7-c650-4e91-aa84-1a66a90076db\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.559193 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1b69f4d7-789b-4992-9b33-a839e3105797-operator-scripts\") pod \"openstack-galera-0\" (UID: \"1b69f4d7-789b-4992-9b33-a839e3105797\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.565976 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/68095fb7-c650-4e91-aa84-1a66a90076db-secrets\") pod \"openstack-galera-1\" (UID: \"68095fb7-c650-4e91-aa84-1a66a90076db\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.566005 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/1b69f4d7-789b-4992-9b33-a839e3105797-secrets\") pod \"openstack-galera-0\" (UID: \"1b69f4d7-789b-4992-9b33-a839e3105797\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.574972 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h54dz\" (UniqueName: \"kubernetes.io/projected/68095fb7-c650-4e91-aa84-1a66a90076db-kube-api-access-h54dz\") pod \"openstack-galera-1\" (UID: \"68095fb7-c650-4e91-aa84-1a66a90076db\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.576924 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"1b69f4d7-789b-4992-9b33-a839e3105797\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.578229 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cz4s\" (UniqueName: \"kubernetes.io/projected/1b69f4d7-789b-4992-9b33-a839e3105797-kube-api-access-5cz4s\") pod \"openstack-galera-0\" (UID: \"1b69f4d7-789b-4992-9b33-a839e3105797\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.579324 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"68095fb7-c650-4e91-aa84-1a66a90076db\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.658909 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/4786b0c4-6846-4873-a3fb-0e155d15e2d0-secrets\") pod \"openstack-galera-2\" (UID: \"4786b0c4-6846-4873-a3fb-0e155d15e2d0\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.658977 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glpv4\" (UniqueName: \"kubernetes.io/projected/4786b0c4-6846-4873-a3fb-0e155d15e2d0-kube-api-access-glpv4\") pod \"openstack-galera-2\" (UID: \"4786b0c4-6846-4873-a3fb-0e155d15e2d0\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.659071 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4786b0c4-6846-4873-a3fb-0e155d15e2d0-config-data-default\") pod \"openstack-galera-2\" (UID: \"4786b0c4-6846-4873-a3fb-0e155d15e2d0\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.659102 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-2\" (UID: \"4786b0c4-6846-4873-a3fb-0e155d15e2d0\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.659150 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4786b0c4-6846-4873-a3fb-0e155d15e2d0-operator-scripts\") pod \"openstack-galera-2\" (UID: \"4786b0c4-6846-4873-a3fb-0e155d15e2d0\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.659205 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4786b0c4-6846-4873-a3fb-0e155d15e2d0-config-data-generated\") pod \"openstack-galera-2\" (UID: \"4786b0c4-6846-4873-a3fb-0e155d15e2d0\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.659229 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4786b0c4-6846-4873-a3fb-0e155d15e2d0-kolla-config\") pod \"openstack-galera-2\" (UID: \"4786b0c4-6846-4873-a3fb-0e155d15e2d0\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.660144 4742 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-2\" (UID: \"4786b0c4-6846-4873-a3fb-0e155d15e2d0\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/openstack-galera-2" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.660335 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4786b0c4-6846-4873-a3fb-0e155d15e2d0-kolla-config\") pod \"openstack-galera-2\" (UID: \"4786b0c4-6846-4873-a3fb-0e155d15e2d0\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.660432 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4786b0c4-6846-4873-a3fb-0e155d15e2d0-config-data-generated\") pod \"openstack-galera-2\" (UID: \"4786b0c4-6846-4873-a3fb-0e155d15e2d0\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.661809 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4786b0c4-6846-4873-a3fb-0e155d15e2d0-config-data-default\") pod \"openstack-galera-2\" (UID: \"4786b0c4-6846-4873-a3fb-0e155d15e2d0\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.661990 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4786b0c4-6846-4873-a3fb-0e155d15e2d0-operator-scripts\") pod \"openstack-galera-2\" (UID: \"4786b0c4-6846-4873-a3fb-0e155d15e2d0\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.663729 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/4786b0c4-6846-4873-a3fb-0e155d15e2d0-secrets\") pod \"openstack-galera-2\" (UID: \"4786b0c4-6846-4873-a3fb-0e155d15e2d0\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.686005 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-2\" (UID: \"4786b0c4-6846-4873-a3fb-0e155d15e2d0\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.691707 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glpv4\" (UniqueName: \"kubernetes.io/projected/4786b0c4-6846-4873-a3fb-0e155d15e2d0-kube-api-access-glpv4\") pod \"openstack-galera-2\" (UID: \"4786b0c4-6846-4873-a3fb-0e155d15e2d0\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.727800 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.752825 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Oct 04 03:23:12 crc kubenswrapper[4742]: I1004 03:23:12.764834 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Oct 04 03:23:13 crc kubenswrapper[4742]: I1004 03:23:13.201816 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-796fff894f-4ns75"] Oct 04 03:23:13 crc kubenswrapper[4742]: I1004 03:23:13.203341 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-796fff894f-4ns75" Oct 04 03:23:13 crc kubenswrapper[4742]: I1004 03:23:13.207350 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-service-cert" Oct 04 03:23:13 crc kubenswrapper[4742]: I1004 03:23:13.207653 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-6tvsg" Oct 04 03:23:13 crc kubenswrapper[4742]: I1004 03:23:13.228642 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-796fff894f-4ns75"] Oct 04 03:23:13 crc kubenswrapper[4742]: I1004 03:23:13.397870 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djscb\" (UniqueName: \"kubernetes.io/projected/533d4297-c693-4bd3-ad07-1a2d59aaa1e1-kube-api-access-djscb\") pod \"infra-operator-controller-manager-796fff894f-4ns75\" (UID: \"533d4297-c693-4bd3-ad07-1a2d59aaa1e1\") " pod="openstack-operators/infra-operator-controller-manager-796fff894f-4ns75" Oct 04 03:23:13 crc kubenswrapper[4742]: I1004 03:23:13.398048 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/533d4297-c693-4bd3-ad07-1a2d59aaa1e1-apiservice-cert\") pod \"infra-operator-controller-manager-796fff894f-4ns75\" (UID: \"533d4297-c693-4bd3-ad07-1a2d59aaa1e1\") " pod="openstack-operators/infra-operator-controller-manager-796fff894f-4ns75" Oct 04 03:23:13 crc kubenswrapper[4742]: I1004 03:23:13.398294 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/533d4297-c693-4bd3-ad07-1a2d59aaa1e1-webhook-cert\") pod \"infra-operator-controller-manager-796fff894f-4ns75\" (UID: \"533d4297-c693-4bd3-ad07-1a2d59aaa1e1\") " pod="openstack-operators/infra-operator-controller-manager-796fff894f-4ns75" Oct 04 03:23:13 crc kubenswrapper[4742]: I1004 03:23:13.458557 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Oct 04 03:23:13 crc kubenswrapper[4742]: W1004 03:23:13.471822 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b69f4d7_789b_4992_9b33_a839e3105797.slice/crio-0d18862cd56fa82a943f99504d51c0de9758d70fb992de2e8140068700662c18 WatchSource:0}: Error finding container 0d18862cd56fa82a943f99504d51c0de9758d70fb992de2e8140068700662c18: Status 404 returned error can't find the container with id 0d18862cd56fa82a943f99504d51c0de9758d70fb992de2e8140068700662c18 Oct 04 03:23:13 crc kubenswrapper[4742]: I1004 03:23:13.500223 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/533d4297-c693-4bd3-ad07-1a2d59aaa1e1-webhook-cert\") pod \"infra-operator-controller-manager-796fff894f-4ns75\" (UID: \"533d4297-c693-4bd3-ad07-1a2d59aaa1e1\") " pod="openstack-operators/infra-operator-controller-manager-796fff894f-4ns75" Oct 04 03:23:13 crc kubenswrapper[4742]: I1004 03:23:13.500859 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djscb\" (UniqueName: \"kubernetes.io/projected/533d4297-c693-4bd3-ad07-1a2d59aaa1e1-kube-api-access-djscb\") pod \"infra-operator-controller-manager-796fff894f-4ns75\" (UID: \"533d4297-c693-4bd3-ad07-1a2d59aaa1e1\") " pod="openstack-operators/infra-operator-controller-manager-796fff894f-4ns75" Oct 04 03:23:13 crc kubenswrapper[4742]: I1004 03:23:13.500895 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/533d4297-c693-4bd3-ad07-1a2d59aaa1e1-apiservice-cert\") pod \"infra-operator-controller-manager-796fff894f-4ns75\" (UID: \"533d4297-c693-4bd3-ad07-1a2d59aaa1e1\") " pod="openstack-operators/infra-operator-controller-manager-796fff894f-4ns75" Oct 04 03:23:13 crc kubenswrapper[4742]: I1004 03:23:13.507254 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/533d4297-c693-4bd3-ad07-1a2d59aaa1e1-apiservice-cert\") pod \"infra-operator-controller-manager-796fff894f-4ns75\" (UID: \"533d4297-c693-4bd3-ad07-1a2d59aaa1e1\") " pod="openstack-operators/infra-operator-controller-manager-796fff894f-4ns75" Oct 04 03:23:13 crc kubenswrapper[4742]: I1004 03:23:13.507254 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/533d4297-c693-4bd3-ad07-1a2d59aaa1e1-webhook-cert\") pod \"infra-operator-controller-manager-796fff894f-4ns75\" (UID: \"533d4297-c693-4bd3-ad07-1a2d59aaa1e1\") " pod="openstack-operators/infra-operator-controller-manager-796fff894f-4ns75" Oct 04 03:23:13 crc kubenswrapper[4742]: I1004 03:23:13.517531 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djscb\" (UniqueName: \"kubernetes.io/projected/533d4297-c693-4bd3-ad07-1a2d59aaa1e1-kube-api-access-djscb\") pod \"infra-operator-controller-manager-796fff894f-4ns75\" (UID: \"533d4297-c693-4bd3-ad07-1a2d59aaa1e1\") " pod="openstack-operators/infra-operator-controller-manager-796fff894f-4ns75" Oct 04 03:23:13 crc kubenswrapper[4742]: I1004 03:23:13.537370 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-796fff894f-4ns75" Oct 04 03:23:13 crc kubenswrapper[4742]: I1004 03:23:13.576953 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Oct 04 03:23:13 crc kubenswrapper[4742]: I1004 03:23:13.676961 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Oct 04 03:23:13 crc kubenswrapper[4742]: I1004 03:23:13.990449 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"68095fb7-c650-4e91-aa84-1a66a90076db","Type":"ContainerStarted","Data":"07dd3cd541aa90fee672b0abec341f298526dd52e3cf59f60c4f2e8475d163d9"} Oct 04 03:23:13 crc kubenswrapper[4742]: I1004 03:23:13.991986 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"4786b0c4-6846-4873-a3fb-0e155d15e2d0","Type":"ContainerStarted","Data":"d1e2edc2bdc5065a84b164d5d896f845d50236513011978195f176841adbd3df"} Oct 04 03:23:13 crc kubenswrapper[4742]: I1004 03:23:13.993041 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"1b69f4d7-789b-4992-9b33-a839e3105797","Type":"ContainerStarted","Data":"0d18862cd56fa82a943f99504d51c0de9758d70fb992de2e8140068700662c18"} Oct 04 03:23:14 crc kubenswrapper[4742]: I1004 03:23:14.049010 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-796fff894f-4ns75"] Oct 04 03:23:14 crc kubenswrapper[4742]: W1004 03:23:14.054033 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod533d4297_c693_4bd3_ad07_1a2d59aaa1e1.slice/crio-277106ac216477b20750ddee29886bcee1ac045367b09120847feb05308c751c WatchSource:0}: Error finding container 277106ac216477b20750ddee29886bcee1ac045367b09120847feb05308c751c: Status 404 returned error can't find the container with id 277106ac216477b20750ddee29886bcee1ac045367b09120847feb05308c751c Oct 04 03:23:15 crc kubenswrapper[4742]: I1004 03:23:15.004013 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-796fff894f-4ns75" event={"ID":"533d4297-c693-4bd3-ad07-1a2d59aaa1e1","Type":"ContainerStarted","Data":"277106ac216477b20750ddee29886bcee1ac045367b09120847feb05308c751c"} Oct 04 03:23:24 crc kubenswrapper[4742]: I1004 03:23:24.083300 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"68095fb7-c650-4e91-aa84-1a66a90076db","Type":"ContainerStarted","Data":"47fa04ab6619cfd4ef12013aa55f04156dbb1a9e6302db6382dd04c99b2deac1"} Oct 04 03:23:24 crc kubenswrapper[4742]: I1004 03:23:24.084757 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"4786b0c4-6846-4873-a3fb-0e155d15e2d0","Type":"ContainerStarted","Data":"2a55a200f2d7df4d7dbbc4de79b60b5500376efa1a3dc95af921a1f41bb03dbb"} Oct 04 03:23:24 crc kubenswrapper[4742]: I1004 03:23:24.086549 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-796fff894f-4ns75" event={"ID":"533d4297-c693-4bd3-ad07-1a2d59aaa1e1","Type":"ContainerStarted","Data":"6e87e0827a2acb49751db652acc91e62d9d8c3383738b9a078fa5334408e4ccc"} Oct 04 03:23:24 crc kubenswrapper[4742]: I1004 03:23:24.086626 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-796fff894f-4ns75" Oct 04 03:23:24 crc kubenswrapper[4742]: I1004 03:23:24.086644 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-796fff894f-4ns75" event={"ID":"533d4297-c693-4bd3-ad07-1a2d59aaa1e1","Type":"ContainerStarted","Data":"ae922efb64f9de8e2e0a8e4a7b70730aa47c4518c0fe4eba6e347a694f780a93"} Oct 04 03:23:24 crc kubenswrapper[4742]: I1004 03:23:24.088461 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"1b69f4d7-789b-4992-9b33-a839e3105797","Type":"ContainerStarted","Data":"48e858ce2c760404f42e3bf75b228aaf1740c8fa0664f3b21a95e169b0ff52d5"} Oct 04 03:23:24 crc kubenswrapper[4742]: I1004 03:23:24.160979 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-796fff894f-4ns75" podStartSLOduration=2.354601843 podStartE2EDuration="11.160954526s" podCreationTimestamp="2025-10-04 03:23:13 +0000 UTC" firstStartedPulling="2025-10-04 03:23:14.063177944 +0000 UTC m=+759.995515120" lastFinishedPulling="2025-10-04 03:23:22.869530577 +0000 UTC m=+768.801867803" observedRunningTime="2025-10-04 03:23:24.154652213 +0000 UTC m=+770.086989389" watchObservedRunningTime="2025-10-04 03:23:24.160954526 +0000 UTC m=+770.093291712" Oct 04 03:23:27 crc kubenswrapper[4742]: I1004 03:23:27.112580 4742 generic.go:334] "Generic (PLEG): container finished" podID="1b69f4d7-789b-4992-9b33-a839e3105797" containerID="48e858ce2c760404f42e3bf75b228aaf1740c8fa0664f3b21a95e169b0ff52d5" exitCode=0 Oct 04 03:23:27 crc kubenswrapper[4742]: I1004 03:23:27.112704 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"1b69f4d7-789b-4992-9b33-a839e3105797","Type":"ContainerDied","Data":"48e858ce2c760404f42e3bf75b228aaf1740c8fa0664f3b21a95e169b0ff52d5"} Oct 04 03:23:27 crc kubenswrapper[4742]: I1004 03:23:27.116961 4742 generic.go:334] "Generic (PLEG): container finished" podID="68095fb7-c650-4e91-aa84-1a66a90076db" containerID="47fa04ab6619cfd4ef12013aa55f04156dbb1a9e6302db6382dd04c99b2deac1" exitCode=0 Oct 04 03:23:27 crc kubenswrapper[4742]: I1004 03:23:27.117023 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"68095fb7-c650-4e91-aa84-1a66a90076db","Type":"ContainerDied","Data":"47fa04ab6619cfd4ef12013aa55f04156dbb1a9e6302db6382dd04c99b2deac1"} Oct 04 03:23:27 crc kubenswrapper[4742]: I1004 03:23:27.121235 4742 generic.go:334] "Generic (PLEG): container finished" podID="4786b0c4-6846-4873-a3fb-0e155d15e2d0" containerID="2a55a200f2d7df4d7dbbc4de79b60b5500376efa1a3dc95af921a1f41bb03dbb" exitCode=0 Oct 04 03:23:27 crc kubenswrapper[4742]: I1004 03:23:27.121356 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"4786b0c4-6846-4873-a3fb-0e155d15e2d0","Type":"ContainerDied","Data":"2a55a200f2d7df4d7dbbc4de79b60b5500376efa1a3dc95af921a1f41bb03dbb"} Oct 04 03:23:28 crc kubenswrapper[4742]: I1004 03:23:28.135856 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"4786b0c4-6846-4873-a3fb-0e155d15e2d0","Type":"ContainerStarted","Data":"3eaedb3a66098a04f862b08cb8365b2755e59e54d101526a9761faa44933d90a"} Oct 04 03:23:28 crc kubenswrapper[4742]: I1004 03:23:28.139660 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"1b69f4d7-789b-4992-9b33-a839e3105797","Type":"ContainerStarted","Data":"58125dd8b6f5073e0166a64d4d68069ebc6975068895d586cf193b987006b8bd"} Oct 04 03:23:28 crc kubenswrapper[4742]: I1004 03:23:28.144169 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"68095fb7-c650-4e91-aa84-1a66a90076db","Type":"ContainerStarted","Data":"10adc10f13a210f0e6bfe35a4832e6367eaa2dcf1bbf394c203fee932bf3af37"} Oct 04 03:23:28 crc kubenswrapper[4742]: I1004 03:23:28.179602 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-2" podStartSLOduration=7.797156175 podStartE2EDuration="17.179571912s" podCreationTimestamp="2025-10-04 03:23:11 +0000 UTC" firstStartedPulling="2025-10-04 03:23:13.587048807 +0000 UTC m=+759.519385983" lastFinishedPulling="2025-10-04 03:23:22.969464504 +0000 UTC m=+768.901801720" observedRunningTime="2025-10-04 03:23:28.173201298 +0000 UTC m=+774.105538534" watchObservedRunningTime="2025-10-04 03:23:28.179571912 +0000 UTC m=+774.111909128" Oct 04 03:23:28 crc kubenswrapper[4742]: I1004 03:23:28.239409 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-1" podStartSLOduration=7.959203254 podStartE2EDuration="17.239382305s" podCreationTimestamp="2025-10-04 03:23:11 +0000 UTC" firstStartedPulling="2025-10-04 03:23:13.736200243 +0000 UTC m=+759.668537419" lastFinishedPulling="2025-10-04 03:23:23.016379294 +0000 UTC m=+768.948716470" observedRunningTime="2025-10-04 03:23:28.236259264 +0000 UTC m=+774.168596510" watchObservedRunningTime="2025-10-04 03:23:28.239382305 +0000 UTC m=+774.171719501" Oct 04 03:23:28 crc kubenswrapper[4742]: I1004 03:23:28.242079 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-0" podStartSLOduration=7.755887941 podStartE2EDuration="17.242063014s" podCreationTimestamp="2025-10-04 03:23:11 +0000 UTC" firstStartedPulling="2025-10-04 03:23:13.473897859 +0000 UTC m=+759.406235025" lastFinishedPulling="2025-10-04 03:23:22.960072912 +0000 UTC m=+768.892410098" observedRunningTime="2025-10-04 03:23:28.211130246 +0000 UTC m=+774.143467452" watchObservedRunningTime="2025-10-04 03:23:28.242063014 +0000 UTC m=+774.174400200" Oct 04 03:23:32 crc kubenswrapper[4742]: I1004 03:23:32.728510 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-0" Oct 04 03:23:32 crc kubenswrapper[4742]: I1004 03:23:32.729320 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-0" Oct 04 03:23:32 crc kubenswrapper[4742]: I1004 03:23:32.753304 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-1" Oct 04 03:23:32 crc kubenswrapper[4742]: I1004 03:23:32.753413 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-1" Oct 04 03:23:32 crc kubenswrapper[4742]: I1004 03:23:32.766581 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-2" Oct 04 03:23:32 crc kubenswrapper[4742]: I1004 03:23:32.766654 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-2" Oct 04 03:23:33 crc kubenswrapper[4742]: I1004 03:23:33.545811 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-796fff894f-4ns75" Oct 04 03:23:36 crc kubenswrapper[4742]: I1004 03:23:36.897006 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-2" Oct 04 03:23:36 crc kubenswrapper[4742]: I1004 03:23:36.968732 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-2" Oct 04 03:23:38 crc kubenswrapper[4742]: I1004 03:23:38.682844 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/memcached-0"] Oct 04 03:23:38 crc kubenswrapper[4742]: I1004 03:23:38.685042 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Oct 04 03:23:38 crc kubenswrapper[4742]: I1004 03:23:38.688751 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"memcached-config-data" Oct 04 03:23:38 crc kubenswrapper[4742]: I1004 03:23:38.689021 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"memcached-memcached-dockercfg-2sk42" Oct 04 03:23:38 crc kubenswrapper[4742]: I1004 03:23:38.696057 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/memcached-0"] Oct 04 03:23:38 crc kubenswrapper[4742]: I1004 03:23:38.801677 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/745e8f2b-0963-4aed-b04e-a21ddc5c9ac7-kolla-config\") pod \"memcached-0\" (UID: \"745e8f2b-0963-4aed-b04e-a21ddc5c9ac7\") " pod="glance-kuttl-tests/memcached-0" Oct 04 03:23:38 crc kubenswrapper[4742]: I1004 03:23:38.801782 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/745e8f2b-0963-4aed-b04e-a21ddc5c9ac7-config-data\") pod \"memcached-0\" (UID: \"745e8f2b-0963-4aed-b04e-a21ddc5c9ac7\") " pod="glance-kuttl-tests/memcached-0" Oct 04 03:23:38 crc kubenswrapper[4742]: I1004 03:23:38.801811 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljnr9\" (UniqueName: \"kubernetes.io/projected/745e8f2b-0963-4aed-b04e-a21ddc5c9ac7-kube-api-access-ljnr9\") pod \"memcached-0\" (UID: \"745e8f2b-0963-4aed-b04e-a21ddc5c9ac7\") " pod="glance-kuttl-tests/memcached-0" Oct 04 03:23:38 crc kubenswrapper[4742]: I1004 03:23:38.902691 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/745e8f2b-0963-4aed-b04e-a21ddc5c9ac7-kolla-config\") pod \"memcached-0\" (UID: \"745e8f2b-0963-4aed-b04e-a21ddc5c9ac7\") " pod="glance-kuttl-tests/memcached-0" Oct 04 03:23:38 crc kubenswrapper[4742]: I1004 03:23:38.902802 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/745e8f2b-0963-4aed-b04e-a21ddc5c9ac7-config-data\") pod \"memcached-0\" (UID: \"745e8f2b-0963-4aed-b04e-a21ddc5c9ac7\") " pod="glance-kuttl-tests/memcached-0" Oct 04 03:23:38 crc kubenswrapper[4742]: I1004 03:23:38.902833 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljnr9\" (UniqueName: \"kubernetes.io/projected/745e8f2b-0963-4aed-b04e-a21ddc5c9ac7-kube-api-access-ljnr9\") pod \"memcached-0\" (UID: \"745e8f2b-0963-4aed-b04e-a21ddc5c9ac7\") " pod="glance-kuttl-tests/memcached-0" Oct 04 03:23:38 crc kubenswrapper[4742]: I1004 03:23:38.903717 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/745e8f2b-0963-4aed-b04e-a21ddc5c9ac7-kolla-config\") pod \"memcached-0\" (UID: \"745e8f2b-0963-4aed-b04e-a21ddc5c9ac7\") " pod="glance-kuttl-tests/memcached-0" Oct 04 03:23:38 crc kubenswrapper[4742]: I1004 03:23:38.904789 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/745e8f2b-0963-4aed-b04e-a21ddc5c9ac7-config-data\") pod \"memcached-0\" (UID: \"745e8f2b-0963-4aed-b04e-a21ddc5c9ac7\") " pod="glance-kuttl-tests/memcached-0" Oct 04 03:23:38 crc kubenswrapper[4742]: I1004 03:23:38.926464 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljnr9\" (UniqueName: \"kubernetes.io/projected/745e8f2b-0963-4aed-b04e-a21ddc5c9ac7-kube-api-access-ljnr9\") pod \"memcached-0\" (UID: \"745e8f2b-0963-4aed-b04e-a21ddc5c9ac7\") " pod="glance-kuttl-tests/memcached-0" Oct 04 03:23:39 crc kubenswrapper[4742]: I1004 03:23:39.002114 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Oct 04 03:23:39 crc kubenswrapper[4742]: I1004 03:23:39.402982 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-m6rpw"] Oct 04 03:23:39 crc kubenswrapper[4742]: I1004 03:23:39.404984 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-m6rpw" Oct 04 03:23:39 crc kubenswrapper[4742]: I1004 03:23:39.408210 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-index-dockercfg-hm649" Oct 04 03:23:39 crc kubenswrapper[4742]: I1004 03:23:39.412968 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-m6rpw"] Oct 04 03:23:39 crc kubenswrapper[4742]: I1004 03:23:39.507540 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/memcached-0"] Oct 04 03:23:39 crc kubenswrapper[4742]: I1004 03:23:39.523978 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vhjm\" (UniqueName: \"kubernetes.io/projected/2cd9293c-3b6d-4a44-a062-a89ca4391588-kube-api-access-4vhjm\") pod \"rabbitmq-cluster-operator-index-m6rpw\" (UID: \"2cd9293c-3b6d-4a44-a062-a89ca4391588\") " pod="openstack-operators/rabbitmq-cluster-operator-index-m6rpw" Oct 04 03:23:39 crc kubenswrapper[4742]: I1004 03:23:39.625567 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vhjm\" (UniqueName: \"kubernetes.io/projected/2cd9293c-3b6d-4a44-a062-a89ca4391588-kube-api-access-4vhjm\") pod \"rabbitmq-cluster-operator-index-m6rpw\" (UID: \"2cd9293c-3b6d-4a44-a062-a89ca4391588\") " pod="openstack-operators/rabbitmq-cluster-operator-index-m6rpw" Oct 04 03:23:39 crc kubenswrapper[4742]: I1004 03:23:39.649565 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vhjm\" (UniqueName: \"kubernetes.io/projected/2cd9293c-3b6d-4a44-a062-a89ca4391588-kube-api-access-4vhjm\") pod \"rabbitmq-cluster-operator-index-m6rpw\" (UID: \"2cd9293c-3b6d-4a44-a062-a89ca4391588\") " pod="openstack-operators/rabbitmq-cluster-operator-index-m6rpw" Oct 04 03:23:39 crc kubenswrapper[4742]: I1004 03:23:39.732306 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-m6rpw" Oct 04 03:23:40 crc kubenswrapper[4742]: I1004 03:23:40.254011 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-m6rpw"] Oct 04 03:23:40 crc kubenswrapper[4742]: W1004 03:23:40.255674 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2cd9293c_3b6d_4a44_a062_a89ca4391588.slice/crio-e84171e50e5b59e962de4e7cc3bb6e5d871688c489465804f0b9660b8142290d WatchSource:0}: Error finding container e84171e50e5b59e962de4e7cc3bb6e5d871688c489465804f0b9660b8142290d: Status 404 returned error can't find the container with id e84171e50e5b59e962de4e7cc3bb6e5d871688c489465804f0b9660b8142290d Oct 04 03:23:40 crc kubenswrapper[4742]: I1004 03:23:40.264929 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"745e8f2b-0963-4aed-b04e-a21ddc5c9ac7","Type":"ContainerStarted","Data":"b732cfc654f232fd53d4abc7af473b4fdaccc02c0670a99a62ce83b7fd367471"} Oct 04 03:23:41 crc kubenswrapper[4742]: I1004 03:23:41.273386 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-m6rpw" event={"ID":"2cd9293c-3b6d-4a44-a062-a89ca4391588","Type":"ContainerStarted","Data":"e84171e50e5b59e962de4e7cc3bb6e5d871688c489465804f0b9660b8142290d"} Oct 04 03:23:42 crc kubenswrapper[4742]: I1004 03:23:42.821921 4742 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/openstack-galera-2" podUID="4786b0c4-6846-4873-a3fb-0e155d15e2d0" containerName="galera" probeResult="failure" output=< Oct 04 03:23:42 crc kubenswrapper[4742]: wsrep_local_state_comment (Donor/Desynced) differs from Synced Oct 04 03:23:42 crc kubenswrapper[4742]: > Oct 04 03:23:43 crc kubenswrapper[4742]: I1004 03:23:43.592405 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-m6rpw"] Oct 04 03:23:44 crc kubenswrapper[4742]: I1004 03:23:44.200320 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-nbgzv"] Oct 04 03:23:44 crc kubenswrapper[4742]: I1004 03:23:44.202220 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-nbgzv" Oct 04 03:23:44 crc kubenswrapper[4742]: I1004 03:23:44.216603 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-nbgzv"] Oct 04 03:23:44 crc kubenswrapper[4742]: I1004 03:23:44.320011 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbfmb\" (UniqueName: \"kubernetes.io/projected/3b222ee4-424b-4962-9722-6ab60c494c6f-kube-api-access-rbfmb\") pod \"rabbitmq-cluster-operator-index-nbgzv\" (UID: \"3b222ee4-424b-4962-9722-6ab60c494c6f\") " pod="openstack-operators/rabbitmq-cluster-operator-index-nbgzv" Oct 04 03:23:44 crc kubenswrapper[4742]: I1004 03:23:44.425130 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbfmb\" (UniqueName: \"kubernetes.io/projected/3b222ee4-424b-4962-9722-6ab60c494c6f-kube-api-access-rbfmb\") pod \"rabbitmq-cluster-operator-index-nbgzv\" (UID: \"3b222ee4-424b-4962-9722-6ab60c494c6f\") " pod="openstack-operators/rabbitmq-cluster-operator-index-nbgzv" Oct 04 03:23:44 crc kubenswrapper[4742]: I1004 03:23:44.450743 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbfmb\" (UniqueName: \"kubernetes.io/projected/3b222ee4-424b-4962-9722-6ab60c494c6f-kube-api-access-rbfmb\") pod \"rabbitmq-cluster-operator-index-nbgzv\" (UID: \"3b222ee4-424b-4962-9722-6ab60c494c6f\") " pod="openstack-operators/rabbitmq-cluster-operator-index-nbgzv" Oct 04 03:23:44 crc kubenswrapper[4742]: I1004 03:23:44.539125 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-nbgzv" Oct 04 03:23:44 crc kubenswrapper[4742]: I1004 03:23:44.872062 4742 patch_prober.go:28] interesting pod/machine-config-daemon-qdt7v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:23:44 crc kubenswrapper[4742]: I1004 03:23:44.872742 4742 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:23:45 crc kubenswrapper[4742]: I1004 03:23:45.528324 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-nbgzv"] Oct 04 03:23:46 crc kubenswrapper[4742]: W1004 03:23:46.309258 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b222ee4_424b_4962_9722_6ab60c494c6f.slice/crio-4bc291f116df0ea470cd0216ee23fa38543e75e0d3ca5ea020169d3bdd29a720 WatchSource:0}: Error finding container 4bc291f116df0ea470cd0216ee23fa38543e75e0d3ca5ea020169d3bdd29a720: Status 404 returned error can't find the container with id 4bc291f116df0ea470cd0216ee23fa38543e75e0d3ca5ea020169d3bdd29a720 Oct 04 03:23:46 crc kubenswrapper[4742]: I1004 03:23:46.337613 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-nbgzv" event={"ID":"3b222ee4-424b-4962-9722-6ab60c494c6f","Type":"ContainerStarted","Data":"4bc291f116df0ea470cd0216ee23fa38543e75e0d3ca5ea020169d3bdd29a720"} Oct 04 03:23:48 crc kubenswrapper[4742]: I1004 03:23:48.355610 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-m6rpw" event={"ID":"2cd9293c-3b6d-4a44-a062-a89ca4391588","Type":"ContainerStarted","Data":"4cdcb8c4dbb3dcf774a1fa115b1d750e9b4f38d8b4d5679c62ebf1b54f0ba7e8"} Oct 04 03:23:48 crc kubenswrapper[4742]: I1004 03:23:48.355828 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/rabbitmq-cluster-operator-index-m6rpw" podUID="2cd9293c-3b6d-4a44-a062-a89ca4391588" containerName="registry-server" containerID="cri-o://4cdcb8c4dbb3dcf774a1fa115b1d750e9b4f38d8b4d5679c62ebf1b54f0ba7e8" gracePeriod=2 Oct 04 03:23:48 crc kubenswrapper[4742]: I1004 03:23:48.358208 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-nbgzv" event={"ID":"3b222ee4-424b-4962-9722-6ab60c494c6f","Type":"ContainerStarted","Data":"4c77dab17cec0320b8097084eda0058941c1b26bb5064cf1b8087d7a69492c1c"} Oct 04 03:23:48 crc kubenswrapper[4742]: I1004 03:23:48.360416 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"745e8f2b-0963-4aed-b04e-a21ddc5c9ac7","Type":"ContainerStarted","Data":"08d01afd67cd890fdf8dfa670b8f061e465d02198d95ea5263d2d4de3040d16e"} Oct 04 03:23:48 crc kubenswrapper[4742]: I1004 03:23:48.360766 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/memcached-0" Oct 04 03:23:48 crc kubenswrapper[4742]: I1004 03:23:48.379719 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-m6rpw" podStartSLOduration=2.235608079 podStartE2EDuration="9.379689907s" podCreationTimestamp="2025-10-04 03:23:39 +0000 UTC" firstStartedPulling="2025-10-04 03:23:40.259651692 +0000 UTC m=+786.191988878" lastFinishedPulling="2025-10-04 03:23:47.40373348 +0000 UTC m=+793.336070706" observedRunningTime="2025-10-04 03:23:48.378824295 +0000 UTC m=+794.311161481" watchObservedRunningTime="2025-10-04 03:23:48.379689907 +0000 UTC m=+794.312027123" Oct 04 03:23:48 crc kubenswrapper[4742]: I1004 03:23:48.452222 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-nbgzv" podStartSLOduration=3.373439252 podStartE2EDuration="4.452184354s" podCreationTimestamp="2025-10-04 03:23:44 +0000 UTC" firstStartedPulling="2025-10-04 03:23:46.323201452 +0000 UTC m=+792.255538638" lastFinishedPulling="2025-10-04 03:23:47.401946564 +0000 UTC m=+793.334283740" observedRunningTime="2025-10-04 03:23:48.444502606 +0000 UTC m=+794.376839792" watchObservedRunningTime="2025-10-04 03:23:48.452184354 +0000 UTC m=+794.384521570" Oct 04 03:23:48 crc kubenswrapper[4742]: I1004 03:23:48.453022 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/memcached-0" podStartSLOduration=5.086886219 podStartE2EDuration="10.453003395s" podCreationTimestamp="2025-10-04 03:23:38 +0000 UTC" firstStartedPulling="2025-10-04 03:23:39.532259778 +0000 UTC m=+785.464596964" lastFinishedPulling="2025-10-04 03:23:44.898376964 +0000 UTC m=+790.830714140" observedRunningTime="2025-10-04 03:23:48.421879133 +0000 UTC m=+794.354216349" watchObservedRunningTime="2025-10-04 03:23:48.453003395 +0000 UTC m=+794.385340611" Oct 04 03:23:48 crc kubenswrapper[4742]: I1004 03:23:48.926334 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-m6rpw" Oct 04 03:23:48 crc kubenswrapper[4742]: I1004 03:23:48.984601 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-1" Oct 04 03:23:49 crc kubenswrapper[4742]: I1004 03:23:49.024228 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-1" Oct 04 03:23:49 crc kubenswrapper[4742]: I1004 03:23:49.071886 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vhjm\" (UniqueName: \"kubernetes.io/projected/2cd9293c-3b6d-4a44-a062-a89ca4391588-kube-api-access-4vhjm\") pod \"2cd9293c-3b6d-4a44-a062-a89ca4391588\" (UID: \"2cd9293c-3b6d-4a44-a062-a89ca4391588\") " Oct 04 03:23:49 crc kubenswrapper[4742]: I1004 03:23:49.079196 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cd9293c-3b6d-4a44-a062-a89ca4391588-kube-api-access-4vhjm" (OuterVolumeSpecName: "kube-api-access-4vhjm") pod "2cd9293c-3b6d-4a44-a062-a89ca4391588" (UID: "2cd9293c-3b6d-4a44-a062-a89ca4391588"). InnerVolumeSpecName "kube-api-access-4vhjm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:23:49 crc kubenswrapper[4742]: I1004 03:23:49.174143 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vhjm\" (UniqueName: \"kubernetes.io/projected/2cd9293c-3b6d-4a44-a062-a89ca4391588-kube-api-access-4vhjm\") on node \"crc\" DevicePath \"\"" Oct 04 03:23:49 crc kubenswrapper[4742]: E1004 03:23:49.222709 4742 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.189:48878->38.102.83.189:41391: write tcp 38.102.83.189:48878->38.102.83.189:41391: write: broken pipe Oct 04 03:23:49 crc kubenswrapper[4742]: I1004 03:23:49.371980 4742 generic.go:334] "Generic (PLEG): container finished" podID="2cd9293c-3b6d-4a44-a062-a89ca4391588" containerID="4cdcb8c4dbb3dcf774a1fa115b1d750e9b4f38d8b4d5679c62ebf1b54f0ba7e8" exitCode=0 Oct 04 03:23:49 crc kubenswrapper[4742]: I1004 03:23:49.372093 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-m6rpw" Oct 04 03:23:49 crc kubenswrapper[4742]: I1004 03:23:49.372166 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-m6rpw" event={"ID":"2cd9293c-3b6d-4a44-a062-a89ca4391588","Type":"ContainerDied","Data":"4cdcb8c4dbb3dcf774a1fa115b1d750e9b4f38d8b4d5679c62ebf1b54f0ba7e8"} Oct 04 03:23:49 crc kubenswrapper[4742]: I1004 03:23:49.372222 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-m6rpw" event={"ID":"2cd9293c-3b6d-4a44-a062-a89ca4391588","Type":"ContainerDied","Data":"e84171e50e5b59e962de4e7cc3bb6e5d871688c489465804f0b9660b8142290d"} Oct 04 03:23:49 crc kubenswrapper[4742]: I1004 03:23:49.372257 4742 scope.go:117] "RemoveContainer" containerID="4cdcb8c4dbb3dcf774a1fa115b1d750e9b4f38d8b4d5679c62ebf1b54f0ba7e8" Oct 04 03:23:49 crc kubenswrapper[4742]: I1004 03:23:49.402976 4742 scope.go:117] "RemoveContainer" containerID="4cdcb8c4dbb3dcf774a1fa115b1d750e9b4f38d8b4d5679c62ebf1b54f0ba7e8" Oct 04 03:23:49 crc kubenswrapper[4742]: E1004 03:23:49.404565 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cdcb8c4dbb3dcf774a1fa115b1d750e9b4f38d8b4d5679c62ebf1b54f0ba7e8\": container with ID starting with 4cdcb8c4dbb3dcf774a1fa115b1d750e9b4f38d8b4d5679c62ebf1b54f0ba7e8 not found: ID does not exist" containerID="4cdcb8c4dbb3dcf774a1fa115b1d750e9b4f38d8b4d5679c62ebf1b54f0ba7e8" Oct 04 03:23:49 crc kubenswrapper[4742]: I1004 03:23:49.404641 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cdcb8c4dbb3dcf774a1fa115b1d750e9b4f38d8b4d5679c62ebf1b54f0ba7e8"} err="failed to get container status \"4cdcb8c4dbb3dcf774a1fa115b1d750e9b4f38d8b4d5679c62ebf1b54f0ba7e8\": rpc error: code = NotFound desc = could not find container \"4cdcb8c4dbb3dcf774a1fa115b1d750e9b4f38d8b4d5679c62ebf1b54f0ba7e8\": container with ID starting with 4cdcb8c4dbb3dcf774a1fa115b1d750e9b4f38d8b4d5679c62ebf1b54f0ba7e8 not found: ID does not exist" Oct 04 03:23:49 crc kubenswrapper[4742]: I1004 03:23:49.426806 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-m6rpw"] Oct 04 03:23:49 crc kubenswrapper[4742]: I1004 03:23:49.433850 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-m6rpw"] Oct 04 03:23:50 crc kubenswrapper[4742]: I1004 03:23:50.873306 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2cd9293c-3b6d-4a44-a062-a89ca4391588" path="/var/lib/kubelet/pods/2cd9293c-3b6d-4a44-a062-a89ca4391588/volumes" Oct 04 03:23:51 crc kubenswrapper[4742]: I1004 03:23:51.419841 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-0" Oct 04 03:23:51 crc kubenswrapper[4742]: I1004 03:23:51.501526 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-0" Oct 04 03:23:54 crc kubenswrapper[4742]: I1004 03:23:54.004208 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/memcached-0" Oct 04 03:23:54 crc kubenswrapper[4742]: I1004 03:23:54.540080 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/rabbitmq-cluster-operator-index-nbgzv" Oct 04 03:23:54 crc kubenswrapper[4742]: I1004 03:23:54.540178 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/rabbitmq-cluster-operator-index-nbgzv" Oct 04 03:23:54 crc kubenswrapper[4742]: I1004 03:23:54.594263 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/rabbitmq-cluster-operator-index-nbgzv" Oct 04 03:23:55 crc kubenswrapper[4742]: I1004 03:23:55.477811 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/rabbitmq-cluster-operator-index-nbgzv" Oct 04 03:23:56 crc kubenswrapper[4742]: I1004 03:23:56.667346 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf"] Oct 04 03:23:56 crc kubenswrapper[4742]: E1004 03:23:56.668261 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cd9293c-3b6d-4a44-a062-a89ca4391588" containerName="registry-server" Oct 04 03:23:56 crc kubenswrapper[4742]: I1004 03:23:56.668312 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cd9293c-3b6d-4a44-a062-a89ca4391588" containerName="registry-server" Oct 04 03:23:56 crc kubenswrapper[4742]: I1004 03:23:56.668561 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cd9293c-3b6d-4a44-a062-a89ca4391588" containerName="registry-server" Oct 04 03:23:56 crc kubenswrapper[4742]: I1004 03:23:56.670144 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf" Oct 04 03:23:56 crc kubenswrapper[4742]: I1004 03:23:56.676163 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-625gw" Oct 04 03:23:56 crc kubenswrapper[4742]: I1004 03:23:56.694341 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf"] Oct 04 03:23:56 crc kubenswrapper[4742]: I1004 03:23:56.701778 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwr9q\" (UniqueName: \"kubernetes.io/projected/7c778fde-b55e-42ae-89bb-90b6396651c4-kube-api-access-bwr9q\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf\" (UID: \"7c778fde-b55e-42ae-89bb-90b6396651c4\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf" Oct 04 03:23:56 crc kubenswrapper[4742]: I1004 03:23:56.701877 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7c778fde-b55e-42ae-89bb-90b6396651c4-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf\" (UID: \"7c778fde-b55e-42ae-89bb-90b6396651c4\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf" Oct 04 03:23:56 crc kubenswrapper[4742]: I1004 03:23:56.702153 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7c778fde-b55e-42ae-89bb-90b6396651c4-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf\" (UID: \"7c778fde-b55e-42ae-89bb-90b6396651c4\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf" Oct 04 03:23:56 crc kubenswrapper[4742]: I1004 03:23:56.804733 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwr9q\" (UniqueName: \"kubernetes.io/projected/7c778fde-b55e-42ae-89bb-90b6396651c4-kube-api-access-bwr9q\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf\" (UID: \"7c778fde-b55e-42ae-89bb-90b6396651c4\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf" Oct 04 03:23:56 crc kubenswrapper[4742]: I1004 03:23:56.804887 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7c778fde-b55e-42ae-89bb-90b6396651c4-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf\" (UID: \"7c778fde-b55e-42ae-89bb-90b6396651c4\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf" Oct 04 03:23:56 crc kubenswrapper[4742]: I1004 03:23:56.804954 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7c778fde-b55e-42ae-89bb-90b6396651c4-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf\" (UID: \"7c778fde-b55e-42ae-89bb-90b6396651c4\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf" Oct 04 03:23:56 crc kubenswrapper[4742]: I1004 03:23:56.805666 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7c778fde-b55e-42ae-89bb-90b6396651c4-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf\" (UID: \"7c778fde-b55e-42ae-89bb-90b6396651c4\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf" Oct 04 03:23:56 crc kubenswrapper[4742]: I1004 03:23:56.805889 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7c778fde-b55e-42ae-89bb-90b6396651c4-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf\" (UID: \"7c778fde-b55e-42ae-89bb-90b6396651c4\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf" Oct 04 03:23:56 crc kubenswrapper[4742]: I1004 03:23:56.853561 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwr9q\" (UniqueName: \"kubernetes.io/projected/7c778fde-b55e-42ae-89bb-90b6396651c4-kube-api-access-bwr9q\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf\" (UID: \"7c778fde-b55e-42ae-89bb-90b6396651c4\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf" Oct 04 03:23:57 crc kubenswrapper[4742]: I1004 03:23:57.002037 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf" Oct 04 03:23:57 crc kubenswrapper[4742]: I1004 03:23:57.481173 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf"] Oct 04 03:23:57 crc kubenswrapper[4742]: W1004 03:23:57.488519 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c778fde_b55e_42ae_89bb_90b6396651c4.slice/crio-dd3925045e049192502c1b0bbe8584c8ee5f30a42ff7d9bee47fb189b880a316 WatchSource:0}: Error finding container dd3925045e049192502c1b0bbe8584c8ee5f30a42ff7d9bee47fb189b880a316: Status 404 returned error can't find the container with id dd3925045e049192502c1b0bbe8584c8ee5f30a42ff7d9bee47fb189b880a316 Oct 04 03:23:58 crc kubenswrapper[4742]: I1004 03:23:58.460914 4742 generic.go:334] "Generic (PLEG): container finished" podID="7c778fde-b55e-42ae-89bb-90b6396651c4" containerID="212c8a1a9d059a5a5de48380ac4721e18ef2a67af1c08573f2e3e64522cee26b" exitCode=0 Oct 04 03:23:58 crc kubenswrapper[4742]: I1004 03:23:58.461047 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf" event={"ID":"7c778fde-b55e-42ae-89bb-90b6396651c4","Type":"ContainerDied","Data":"212c8a1a9d059a5a5de48380ac4721e18ef2a67af1c08573f2e3e64522cee26b"} Oct 04 03:23:58 crc kubenswrapper[4742]: I1004 03:23:58.461549 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf" event={"ID":"7c778fde-b55e-42ae-89bb-90b6396651c4","Type":"ContainerStarted","Data":"dd3925045e049192502c1b0bbe8584c8ee5f30a42ff7d9bee47fb189b880a316"} Oct 04 03:24:00 crc kubenswrapper[4742]: I1004 03:24:00.481981 4742 generic.go:334] "Generic (PLEG): container finished" podID="7c778fde-b55e-42ae-89bb-90b6396651c4" containerID="1b272c5ce4c91675238681d6875a54d2e89f91306b88f3c52e0cbeb017303934" exitCode=0 Oct 04 03:24:00 crc kubenswrapper[4742]: I1004 03:24:00.482540 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf" event={"ID":"7c778fde-b55e-42ae-89bb-90b6396651c4","Type":"ContainerDied","Data":"1b272c5ce4c91675238681d6875a54d2e89f91306b88f3c52e0cbeb017303934"} Oct 04 03:24:01 crc kubenswrapper[4742]: I1004 03:24:01.495698 4742 generic.go:334] "Generic (PLEG): container finished" podID="7c778fde-b55e-42ae-89bb-90b6396651c4" containerID="9cf1f154ec4708d67bd7b3205a2eab2a7a0d9c4692adfe0239d915cc419be541" exitCode=0 Oct 04 03:24:01 crc kubenswrapper[4742]: I1004 03:24:01.495823 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf" event={"ID":"7c778fde-b55e-42ae-89bb-90b6396651c4","Type":"ContainerDied","Data":"9cf1f154ec4708d67bd7b3205a2eab2a7a0d9c4692adfe0239d915cc419be541"} Oct 04 03:24:02 crc kubenswrapper[4742]: I1004 03:24:02.965117 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf" Oct 04 03:24:03 crc kubenswrapper[4742]: I1004 03:24:03.011959 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7c778fde-b55e-42ae-89bb-90b6396651c4-bundle\") pod \"7c778fde-b55e-42ae-89bb-90b6396651c4\" (UID: \"7c778fde-b55e-42ae-89bb-90b6396651c4\") " Oct 04 03:24:03 crc kubenswrapper[4742]: I1004 03:24:03.012053 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7c778fde-b55e-42ae-89bb-90b6396651c4-util\") pod \"7c778fde-b55e-42ae-89bb-90b6396651c4\" (UID: \"7c778fde-b55e-42ae-89bb-90b6396651c4\") " Oct 04 03:24:03 crc kubenswrapper[4742]: I1004 03:24:03.012130 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwr9q\" (UniqueName: \"kubernetes.io/projected/7c778fde-b55e-42ae-89bb-90b6396651c4-kube-api-access-bwr9q\") pod \"7c778fde-b55e-42ae-89bb-90b6396651c4\" (UID: \"7c778fde-b55e-42ae-89bb-90b6396651c4\") " Oct 04 03:24:03 crc kubenswrapper[4742]: I1004 03:24:03.013225 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c778fde-b55e-42ae-89bb-90b6396651c4-bundle" (OuterVolumeSpecName: "bundle") pod "7c778fde-b55e-42ae-89bb-90b6396651c4" (UID: "7c778fde-b55e-42ae-89bb-90b6396651c4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:24:03 crc kubenswrapper[4742]: I1004 03:24:03.023246 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c778fde-b55e-42ae-89bb-90b6396651c4-kube-api-access-bwr9q" (OuterVolumeSpecName: "kube-api-access-bwr9q") pod "7c778fde-b55e-42ae-89bb-90b6396651c4" (UID: "7c778fde-b55e-42ae-89bb-90b6396651c4"). InnerVolumeSpecName "kube-api-access-bwr9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:24:03 crc kubenswrapper[4742]: I1004 03:24:03.040582 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c778fde-b55e-42ae-89bb-90b6396651c4-util" (OuterVolumeSpecName: "util") pod "7c778fde-b55e-42ae-89bb-90b6396651c4" (UID: "7c778fde-b55e-42ae-89bb-90b6396651c4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:24:03 crc kubenswrapper[4742]: I1004 03:24:03.113983 4742 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7c778fde-b55e-42ae-89bb-90b6396651c4-util\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:03 crc kubenswrapper[4742]: I1004 03:24:03.114042 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwr9q\" (UniqueName: \"kubernetes.io/projected/7c778fde-b55e-42ae-89bb-90b6396651c4-kube-api-access-bwr9q\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:03 crc kubenswrapper[4742]: I1004 03:24:03.114065 4742 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7c778fde-b55e-42ae-89bb-90b6396651c4-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:03 crc kubenswrapper[4742]: I1004 03:24:03.516751 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf" event={"ID":"7c778fde-b55e-42ae-89bb-90b6396651c4","Type":"ContainerDied","Data":"dd3925045e049192502c1b0bbe8584c8ee5f30a42ff7d9bee47fb189b880a316"} Oct 04 03:24:03 crc kubenswrapper[4742]: I1004 03:24:03.517500 4742 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd3925045e049192502c1b0bbe8584c8ee5f30a42ff7d9bee47fb189b880a316" Oct 04 03:24:03 crc kubenswrapper[4742]: I1004 03:24:03.516847 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf" Oct 04 03:24:11 crc kubenswrapper[4742]: I1004 03:24:11.344982 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-8c4x6"] Oct 04 03:24:11 crc kubenswrapper[4742]: E1004 03:24:11.348717 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c778fde-b55e-42ae-89bb-90b6396651c4" containerName="extract" Oct 04 03:24:11 crc kubenswrapper[4742]: I1004 03:24:11.348744 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c778fde-b55e-42ae-89bb-90b6396651c4" containerName="extract" Oct 04 03:24:11 crc kubenswrapper[4742]: E1004 03:24:11.348772 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c778fde-b55e-42ae-89bb-90b6396651c4" containerName="pull" Oct 04 03:24:11 crc kubenswrapper[4742]: I1004 03:24:11.348785 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c778fde-b55e-42ae-89bb-90b6396651c4" containerName="pull" Oct 04 03:24:11 crc kubenswrapper[4742]: E1004 03:24:11.348801 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c778fde-b55e-42ae-89bb-90b6396651c4" containerName="util" Oct 04 03:24:11 crc kubenswrapper[4742]: I1004 03:24:11.348813 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c778fde-b55e-42ae-89bb-90b6396651c4" containerName="util" Oct 04 03:24:11 crc kubenswrapper[4742]: I1004 03:24:11.349044 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c778fde-b55e-42ae-89bb-90b6396651c4" containerName="extract" Oct 04 03:24:11 crc kubenswrapper[4742]: I1004 03:24:11.349772 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-8c4x6" Oct 04 03:24:11 crc kubenswrapper[4742]: I1004 03:24:11.355854 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-8c4x6"] Oct 04 03:24:11 crc kubenswrapper[4742]: I1004 03:24:11.358773 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-dockercfg-78fj8" Oct 04 03:24:11 crc kubenswrapper[4742]: I1004 03:24:11.454512 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s64g4\" (UniqueName: \"kubernetes.io/projected/b846aad4-0d5d-4023-bed1-59da3710e48a-kube-api-access-s64g4\") pod \"rabbitmq-cluster-operator-779fc9694b-8c4x6\" (UID: \"b846aad4-0d5d-4023-bed1-59da3710e48a\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-8c4x6" Oct 04 03:24:11 crc kubenswrapper[4742]: I1004 03:24:11.556812 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s64g4\" (UniqueName: \"kubernetes.io/projected/b846aad4-0d5d-4023-bed1-59da3710e48a-kube-api-access-s64g4\") pod \"rabbitmq-cluster-operator-779fc9694b-8c4x6\" (UID: \"b846aad4-0d5d-4023-bed1-59da3710e48a\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-8c4x6" Oct 04 03:24:11 crc kubenswrapper[4742]: I1004 03:24:11.587622 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s64g4\" (UniqueName: \"kubernetes.io/projected/b846aad4-0d5d-4023-bed1-59da3710e48a-kube-api-access-s64g4\") pod \"rabbitmq-cluster-operator-779fc9694b-8c4x6\" (UID: \"b846aad4-0d5d-4023-bed1-59da3710e48a\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-8c4x6" Oct 04 03:24:11 crc kubenswrapper[4742]: I1004 03:24:11.674365 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-8c4x6" Oct 04 03:24:12 crc kubenswrapper[4742]: I1004 03:24:12.155045 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-8c4x6"] Oct 04 03:24:12 crc kubenswrapper[4742]: W1004 03:24:12.168444 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podb846aad4_0d5d_4023_bed1_59da3710e48a.slice/crio-35a91c75cb1fb1982f9f6f2c39c81e3d624ca0dda62f8266a5916a46401e94eb WatchSource:0}: Error finding container 35a91c75cb1fb1982f9f6f2c39c81e3d624ca0dda62f8266a5916a46401e94eb: Status 404 returned error can't find the container with id 35a91c75cb1fb1982f9f6f2c39c81e3d624ca0dda62f8266a5916a46401e94eb Oct 04 03:24:12 crc kubenswrapper[4742]: I1004 03:24:12.589925 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-8c4x6" event={"ID":"b846aad4-0d5d-4023-bed1-59da3710e48a","Type":"ContainerStarted","Data":"35a91c75cb1fb1982f9f6f2c39c81e3d624ca0dda62f8266a5916a46401e94eb"} Oct 04 03:24:14 crc kubenswrapper[4742]: I1004 03:24:14.615876 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-8c4x6" event={"ID":"b846aad4-0d5d-4023-bed1-59da3710e48a","Type":"ContainerStarted","Data":"fdd648c65cdd605a0b32bfaa6e0c326348494e989d3d5f1cf00d34f86796b006"} Oct 04 03:24:14 crc kubenswrapper[4742]: I1004 03:24:14.639871 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-8c4x6" podStartSLOduration=1.7574650840000001 podStartE2EDuration="3.639846016s" podCreationTimestamp="2025-10-04 03:24:11 +0000 UTC" firstStartedPulling="2025-10-04 03:24:12.171176454 +0000 UTC m=+818.103513630" lastFinishedPulling="2025-10-04 03:24:14.053557386 +0000 UTC m=+819.985894562" observedRunningTime="2025-10-04 03:24:14.635416272 +0000 UTC m=+820.567753458" watchObservedRunningTime="2025-10-04 03:24:14.639846016 +0000 UTC m=+820.572183202" Oct 04 03:24:14 crc kubenswrapper[4742]: I1004 03:24:14.912085 4742 patch_prober.go:28] interesting pod/machine-config-daemon-qdt7v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:24:14 crc kubenswrapper[4742]: I1004 03:24:14.912198 4742 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:24:21 crc kubenswrapper[4742]: I1004 03:24:21.150173 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Oct 04 03:24:21 crc kubenswrapper[4742]: I1004 03:24:21.152866 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 03:24:21 crc kubenswrapper[4742]: I1004 03:24:21.155609 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-erlang-cookie" Oct 04 03:24:21 crc kubenswrapper[4742]: I1004 03:24:21.155898 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-default-user" Oct 04 03:24:21 crc kubenswrapper[4742]: I1004 03:24:21.158317 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-server-dockercfg-ndfk8" Oct 04 03:24:21 crc kubenswrapper[4742]: I1004 03:24:21.158352 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-server-conf" Oct 04 03:24:21 crc kubenswrapper[4742]: I1004 03:24:21.158405 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-plugins-conf" Oct 04 03:24:21 crc kubenswrapper[4742]: I1004 03:24:21.167151 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Oct 04 03:24:21 crc kubenswrapper[4742]: I1004 03:24:21.225006 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8e6c595f-148d-419e-8a72-a991c62df809-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8e6c595f-148d-419e-8a72-a991c62df809\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 03:24:21 crc kubenswrapper[4742]: I1004 03:24:21.225048 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8e6c595f-148d-419e-8a72-a991c62df809-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8e6c595f-148d-419e-8a72-a991c62df809\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 03:24:21 crc kubenswrapper[4742]: I1004 03:24:21.225097 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6trl\" (UniqueName: \"kubernetes.io/projected/8e6c595f-148d-419e-8a72-a991c62df809-kube-api-access-x6trl\") pod \"rabbitmq-server-0\" (UID: \"8e6c595f-148d-419e-8a72-a991c62df809\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 03:24:21 crc kubenswrapper[4742]: I1004 03:24:21.225125 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8e6c595f-148d-419e-8a72-a991c62df809-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8e6c595f-148d-419e-8a72-a991c62df809\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 03:24:21 crc kubenswrapper[4742]: I1004 03:24:21.225159 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-9d154609-0a0f-41aa-882d-7a79d56adfa5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9d154609-0a0f-41aa-882d-7a79d56adfa5\") pod \"rabbitmq-server-0\" (UID: \"8e6c595f-148d-419e-8a72-a991c62df809\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 03:24:21 crc kubenswrapper[4742]: I1004 03:24:21.225183 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8e6c595f-148d-419e-8a72-a991c62df809-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8e6c595f-148d-419e-8a72-a991c62df809\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 03:24:21 crc kubenswrapper[4742]: I1004 03:24:21.225219 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8e6c595f-148d-419e-8a72-a991c62df809-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8e6c595f-148d-419e-8a72-a991c62df809\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 03:24:21 crc kubenswrapper[4742]: I1004 03:24:21.225251 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8e6c595f-148d-419e-8a72-a991c62df809-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8e6c595f-148d-419e-8a72-a991c62df809\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 03:24:21 crc kubenswrapper[4742]: I1004 03:24:21.326476 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8e6c595f-148d-419e-8a72-a991c62df809-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8e6c595f-148d-419e-8a72-a991c62df809\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 03:24:21 crc kubenswrapper[4742]: I1004 03:24:21.326529 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8e6c595f-148d-419e-8a72-a991c62df809-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8e6c595f-148d-419e-8a72-a991c62df809\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 03:24:21 crc kubenswrapper[4742]: I1004 03:24:21.326591 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6trl\" (UniqueName: \"kubernetes.io/projected/8e6c595f-148d-419e-8a72-a991c62df809-kube-api-access-x6trl\") pod \"rabbitmq-server-0\" (UID: \"8e6c595f-148d-419e-8a72-a991c62df809\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 03:24:21 crc kubenswrapper[4742]: I1004 03:24:21.326617 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8e6c595f-148d-419e-8a72-a991c62df809-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8e6c595f-148d-419e-8a72-a991c62df809\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 03:24:21 crc kubenswrapper[4742]: I1004 03:24:21.326656 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-9d154609-0a0f-41aa-882d-7a79d56adfa5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9d154609-0a0f-41aa-882d-7a79d56adfa5\") pod \"rabbitmq-server-0\" (UID: \"8e6c595f-148d-419e-8a72-a991c62df809\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 03:24:21 crc kubenswrapper[4742]: I1004 03:24:21.326684 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8e6c595f-148d-419e-8a72-a991c62df809-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8e6c595f-148d-419e-8a72-a991c62df809\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 03:24:21 crc kubenswrapper[4742]: I1004 03:24:21.326723 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8e6c595f-148d-419e-8a72-a991c62df809-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8e6c595f-148d-419e-8a72-a991c62df809\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 03:24:21 crc kubenswrapper[4742]: I1004 03:24:21.326755 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8e6c595f-148d-419e-8a72-a991c62df809-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8e6c595f-148d-419e-8a72-a991c62df809\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 03:24:21 crc kubenswrapper[4742]: I1004 03:24:21.327710 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8e6c595f-148d-419e-8a72-a991c62df809-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8e6c595f-148d-419e-8a72-a991c62df809\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 03:24:21 crc kubenswrapper[4742]: I1004 03:24:21.328507 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8e6c595f-148d-419e-8a72-a991c62df809-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8e6c595f-148d-419e-8a72-a991c62df809\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 03:24:21 crc kubenswrapper[4742]: I1004 03:24:21.328635 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8e6c595f-148d-419e-8a72-a991c62df809-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8e6c595f-148d-419e-8a72-a991c62df809\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 03:24:21 crc kubenswrapper[4742]: I1004 03:24:21.332158 4742 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 04 03:24:21 crc kubenswrapper[4742]: I1004 03:24:21.332196 4742 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-9d154609-0a0f-41aa-882d-7a79d56adfa5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9d154609-0a0f-41aa-882d-7a79d56adfa5\") pod \"rabbitmq-server-0\" (UID: \"8e6c595f-148d-419e-8a72-a991c62df809\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/263c02ae89154d87ef1682d646109bd17d07b1b1adb070dee977b72d1219e0b2/globalmount\"" pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 03:24:21 crc kubenswrapper[4742]: I1004 03:24:21.337840 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8e6c595f-148d-419e-8a72-a991c62df809-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8e6c595f-148d-419e-8a72-a991c62df809\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 03:24:21 crc kubenswrapper[4742]: I1004 03:24:21.338805 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8e6c595f-148d-419e-8a72-a991c62df809-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8e6c595f-148d-419e-8a72-a991c62df809\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 03:24:21 crc kubenswrapper[4742]: I1004 03:24:21.344944 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8e6c595f-148d-419e-8a72-a991c62df809-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8e6c595f-148d-419e-8a72-a991c62df809\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 03:24:21 crc kubenswrapper[4742]: I1004 03:24:21.348097 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6trl\" (UniqueName: \"kubernetes.io/projected/8e6c595f-148d-419e-8a72-a991c62df809-kube-api-access-x6trl\") pod \"rabbitmq-server-0\" (UID: \"8e6c595f-148d-419e-8a72-a991c62df809\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 03:24:21 crc kubenswrapper[4742]: I1004 03:24:21.369077 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-9d154609-0a0f-41aa-882d-7a79d56adfa5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9d154609-0a0f-41aa-882d-7a79d56adfa5\") pod \"rabbitmq-server-0\" (UID: \"8e6c595f-148d-419e-8a72-a991c62df809\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 03:24:21 crc kubenswrapper[4742]: I1004 03:24:21.510646 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 03:24:21 crc kubenswrapper[4742]: I1004 03:24:21.787756 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Oct 04 03:24:22 crc kubenswrapper[4742]: I1004 03:24:22.708492 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"8e6c595f-148d-419e-8a72-a991c62df809","Type":"ContainerStarted","Data":"6af7c8017845778ece79d70b66410ad2d8a3de6071baf81b513bfcf42dfa0eed"} Oct 04 03:24:23 crc kubenswrapper[4742]: I1004 03:24:23.202115 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-kbkhk"] Oct 04 03:24:23 crc kubenswrapper[4742]: I1004 03:24:23.202900 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-kbkhk" Oct 04 03:24:23 crc kubenswrapper[4742]: I1004 03:24:23.205614 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-index-dockercfg-vxgfl" Oct 04 03:24:23 crc kubenswrapper[4742]: I1004 03:24:23.215301 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-kbkhk"] Oct 04 03:24:23 crc kubenswrapper[4742]: I1004 03:24:23.365319 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qf4ws\" (UniqueName: \"kubernetes.io/projected/ca9efb6f-8e2d-43bc-ac51-a8673024fe54-kube-api-access-qf4ws\") pod \"keystone-operator-index-kbkhk\" (UID: \"ca9efb6f-8e2d-43bc-ac51-a8673024fe54\") " pod="openstack-operators/keystone-operator-index-kbkhk" Oct 04 03:24:23 crc kubenswrapper[4742]: I1004 03:24:23.467521 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qf4ws\" (UniqueName: \"kubernetes.io/projected/ca9efb6f-8e2d-43bc-ac51-a8673024fe54-kube-api-access-qf4ws\") pod \"keystone-operator-index-kbkhk\" (UID: \"ca9efb6f-8e2d-43bc-ac51-a8673024fe54\") " pod="openstack-operators/keystone-operator-index-kbkhk" Oct 04 03:24:23 crc kubenswrapper[4742]: I1004 03:24:23.499384 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qf4ws\" (UniqueName: \"kubernetes.io/projected/ca9efb6f-8e2d-43bc-ac51-a8673024fe54-kube-api-access-qf4ws\") pod \"keystone-operator-index-kbkhk\" (UID: \"ca9efb6f-8e2d-43bc-ac51-a8673024fe54\") " pod="openstack-operators/keystone-operator-index-kbkhk" Oct 04 03:24:23 crc kubenswrapper[4742]: I1004 03:24:23.542452 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-kbkhk" Oct 04 03:24:25 crc kubenswrapper[4742]: I1004 03:24:25.815350 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-58ppm"] Oct 04 03:24:25 crc kubenswrapper[4742]: I1004 03:24:25.817926 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-58ppm" Oct 04 03:24:25 crc kubenswrapper[4742]: I1004 03:24:25.833852 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-58ppm"] Oct 04 03:24:25 crc kubenswrapper[4742]: I1004 03:24:25.920483 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/125f69b5-d78b-4241-b3d6-c1a151d567e0-catalog-content\") pod \"certified-operators-58ppm\" (UID: \"125f69b5-d78b-4241-b3d6-c1a151d567e0\") " pod="openshift-marketplace/certified-operators-58ppm" Oct 04 03:24:25 crc kubenswrapper[4742]: I1004 03:24:25.920643 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvb6k\" (UniqueName: \"kubernetes.io/projected/125f69b5-d78b-4241-b3d6-c1a151d567e0-kube-api-access-vvb6k\") pod \"certified-operators-58ppm\" (UID: \"125f69b5-d78b-4241-b3d6-c1a151d567e0\") " pod="openshift-marketplace/certified-operators-58ppm" Oct 04 03:24:25 crc kubenswrapper[4742]: I1004 03:24:25.920676 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/125f69b5-d78b-4241-b3d6-c1a151d567e0-utilities\") pod \"certified-operators-58ppm\" (UID: \"125f69b5-d78b-4241-b3d6-c1a151d567e0\") " pod="openshift-marketplace/certified-operators-58ppm" Oct 04 03:24:26 crc kubenswrapper[4742]: I1004 03:24:26.021869 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvb6k\" (UniqueName: \"kubernetes.io/projected/125f69b5-d78b-4241-b3d6-c1a151d567e0-kube-api-access-vvb6k\") pod \"certified-operators-58ppm\" (UID: \"125f69b5-d78b-4241-b3d6-c1a151d567e0\") " pod="openshift-marketplace/certified-operators-58ppm" Oct 04 03:24:26 crc kubenswrapper[4742]: I1004 03:24:26.022409 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/125f69b5-d78b-4241-b3d6-c1a151d567e0-utilities\") pod \"certified-operators-58ppm\" (UID: \"125f69b5-d78b-4241-b3d6-c1a151d567e0\") " pod="openshift-marketplace/certified-operators-58ppm" Oct 04 03:24:26 crc kubenswrapper[4742]: I1004 03:24:26.022524 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/125f69b5-d78b-4241-b3d6-c1a151d567e0-catalog-content\") pod \"certified-operators-58ppm\" (UID: \"125f69b5-d78b-4241-b3d6-c1a151d567e0\") " pod="openshift-marketplace/certified-operators-58ppm" Oct 04 03:24:26 crc kubenswrapper[4742]: I1004 03:24:26.023110 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/125f69b5-d78b-4241-b3d6-c1a151d567e0-catalog-content\") pod \"certified-operators-58ppm\" (UID: \"125f69b5-d78b-4241-b3d6-c1a151d567e0\") " pod="openshift-marketplace/certified-operators-58ppm" Oct 04 03:24:26 crc kubenswrapper[4742]: I1004 03:24:26.023209 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/125f69b5-d78b-4241-b3d6-c1a151d567e0-utilities\") pod \"certified-operators-58ppm\" (UID: \"125f69b5-d78b-4241-b3d6-c1a151d567e0\") " pod="openshift-marketplace/certified-operators-58ppm" Oct 04 03:24:26 crc kubenswrapper[4742]: I1004 03:24:26.047118 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvb6k\" (UniqueName: \"kubernetes.io/projected/125f69b5-d78b-4241-b3d6-c1a151d567e0-kube-api-access-vvb6k\") pod \"certified-operators-58ppm\" (UID: \"125f69b5-d78b-4241-b3d6-c1a151d567e0\") " pod="openshift-marketplace/certified-operators-58ppm" Oct 04 03:24:26 crc kubenswrapper[4742]: I1004 03:24:26.141325 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-58ppm" Oct 04 03:24:26 crc kubenswrapper[4742]: I1004 03:24:26.623755 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-kbkhk"] Oct 04 03:24:27 crc kubenswrapper[4742]: W1004 03:24:27.451650 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca9efb6f_8e2d_43bc_ac51_a8673024fe54.slice/crio-9760f566ef11c1ee2aa71d1f58a7be53b7293e8f9490bca4d683e687db7d226b WatchSource:0}: Error finding container 9760f566ef11c1ee2aa71d1f58a7be53b7293e8f9490bca4d683e687db7d226b: Status 404 returned error can't find the container with id 9760f566ef11c1ee2aa71d1f58a7be53b7293e8f9490bca4d683e687db7d226b Oct 04 03:24:27 crc kubenswrapper[4742]: I1004 03:24:27.772933 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-kbkhk" event={"ID":"ca9efb6f-8e2d-43bc-ac51-a8673024fe54","Type":"ContainerStarted","Data":"9760f566ef11c1ee2aa71d1f58a7be53b7293e8f9490bca4d683e687db7d226b"} Oct 04 03:24:28 crc kubenswrapper[4742]: I1004 03:24:28.807385 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-58ppm"] Oct 04 03:24:28 crc kubenswrapper[4742]: W1004 03:24:28.825499 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod125f69b5_d78b_4241_b3d6_c1a151d567e0.slice/crio-a7456aa97f293c6ff84e50a9a3d9ec289835cef7eda10e2b8035a9fa10b64cf7 WatchSource:0}: Error finding container a7456aa97f293c6ff84e50a9a3d9ec289835cef7eda10e2b8035a9fa10b64cf7: Status 404 returned error can't find the container with id a7456aa97f293c6ff84e50a9a3d9ec289835cef7eda10e2b8035a9fa10b64cf7 Oct 04 03:24:29 crc kubenswrapper[4742]: I1004 03:24:29.810427 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-kbkhk" event={"ID":"ca9efb6f-8e2d-43bc-ac51-a8673024fe54","Type":"ContainerStarted","Data":"7f626857c61c54a9a53c98efed81029e8222ca0fcbb46359ef11fd94fe02e5b9"} Oct 04 03:24:29 crc kubenswrapper[4742]: I1004 03:24:29.811903 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"8e6c595f-148d-419e-8a72-a991c62df809","Type":"ContainerStarted","Data":"de3e9b1fa72f15a5970dc89363224774e34de58289ec158a7add75fd811b8d23"} Oct 04 03:24:29 crc kubenswrapper[4742]: I1004 03:24:29.813712 4742 generic.go:334] "Generic (PLEG): container finished" podID="125f69b5-d78b-4241-b3d6-c1a151d567e0" containerID="21c3ca01c55e96e1e5315b61c3f48794019088830fcd21656248adfbb05221e5" exitCode=0 Oct 04 03:24:29 crc kubenswrapper[4742]: I1004 03:24:29.813742 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-58ppm" event={"ID":"125f69b5-d78b-4241-b3d6-c1a151d567e0","Type":"ContainerDied","Data":"21c3ca01c55e96e1e5315b61c3f48794019088830fcd21656248adfbb05221e5"} Oct 04 03:24:29 crc kubenswrapper[4742]: I1004 03:24:29.813878 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-58ppm" event={"ID":"125f69b5-d78b-4241-b3d6-c1a151d567e0","Type":"ContainerStarted","Data":"a7456aa97f293c6ff84e50a9a3d9ec289835cef7eda10e2b8035a9fa10b64cf7"} Oct 04 03:24:29 crc kubenswrapper[4742]: I1004 03:24:29.877327 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-kbkhk" podStartSLOduration=5.59013411 podStartE2EDuration="6.877310402s" podCreationTimestamp="2025-10-04 03:24:23 +0000 UTC" firstStartedPulling="2025-10-04 03:24:27.456817222 +0000 UTC m=+833.389154398" lastFinishedPulling="2025-10-04 03:24:28.743993504 +0000 UTC m=+834.676330690" observedRunningTime="2025-10-04 03:24:29.834317745 +0000 UTC m=+835.766654941" watchObservedRunningTime="2025-10-04 03:24:29.877310402 +0000 UTC m=+835.809647578" Oct 04 03:24:30 crc kubenswrapper[4742]: I1004 03:24:30.825453 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-58ppm" event={"ID":"125f69b5-d78b-4241-b3d6-c1a151d567e0","Type":"ContainerStarted","Data":"b2de163d5213a3ba66c6b26e9140095000c49edf514825393252e1799927bd56"} Oct 04 03:24:31 crc kubenswrapper[4742]: I1004 03:24:31.835745 4742 generic.go:334] "Generic (PLEG): container finished" podID="125f69b5-d78b-4241-b3d6-c1a151d567e0" containerID="b2de163d5213a3ba66c6b26e9140095000c49edf514825393252e1799927bd56" exitCode=0 Oct 04 03:24:31 crc kubenswrapper[4742]: I1004 03:24:31.835813 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-58ppm" event={"ID":"125f69b5-d78b-4241-b3d6-c1a151d567e0","Type":"ContainerDied","Data":"b2de163d5213a3ba66c6b26e9140095000c49edf514825393252e1799927bd56"} Oct 04 03:24:32 crc kubenswrapper[4742]: I1004 03:24:32.844746 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-58ppm" event={"ID":"125f69b5-d78b-4241-b3d6-c1a151d567e0","Type":"ContainerStarted","Data":"96fc5bf80a98cd07699381311b67bdd7694e7ce11da0c6786c54647ae556e981"} Oct 04 03:24:32 crc kubenswrapper[4742]: I1004 03:24:32.879645 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-58ppm" podStartSLOduration=5.289021226 podStartE2EDuration="7.879618908s" podCreationTimestamp="2025-10-04 03:24:25 +0000 UTC" firstStartedPulling="2025-10-04 03:24:29.814916515 +0000 UTC m=+835.747253691" lastFinishedPulling="2025-10-04 03:24:32.405514167 +0000 UTC m=+838.337851373" observedRunningTime="2025-10-04 03:24:32.878810027 +0000 UTC m=+838.811147203" watchObservedRunningTime="2025-10-04 03:24:32.879618908 +0000 UTC m=+838.811956074" Oct 04 03:24:33 crc kubenswrapper[4742]: I1004 03:24:33.415798 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dqltg"] Oct 04 03:24:33 crc kubenswrapper[4742]: I1004 03:24:33.417379 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dqltg" Oct 04 03:24:33 crc kubenswrapper[4742]: I1004 03:24:33.432903 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dqltg"] Oct 04 03:24:33 crc kubenswrapper[4742]: I1004 03:24:33.542671 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-index-kbkhk" Oct 04 03:24:33 crc kubenswrapper[4742]: I1004 03:24:33.542742 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/keystone-operator-index-kbkhk" Oct 04 03:24:33 crc kubenswrapper[4742]: I1004 03:24:33.558947 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90332c58-6092-4a1d-b84f-27ab854928f7-catalog-content\") pod \"redhat-operators-dqltg\" (UID: \"90332c58-6092-4a1d-b84f-27ab854928f7\") " pod="openshift-marketplace/redhat-operators-dqltg" Oct 04 03:24:33 crc kubenswrapper[4742]: I1004 03:24:33.559030 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbt2n\" (UniqueName: \"kubernetes.io/projected/90332c58-6092-4a1d-b84f-27ab854928f7-kube-api-access-nbt2n\") pod \"redhat-operators-dqltg\" (UID: \"90332c58-6092-4a1d-b84f-27ab854928f7\") " pod="openshift-marketplace/redhat-operators-dqltg" Oct 04 03:24:33 crc kubenswrapper[4742]: I1004 03:24:33.559058 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90332c58-6092-4a1d-b84f-27ab854928f7-utilities\") pod \"redhat-operators-dqltg\" (UID: \"90332c58-6092-4a1d-b84f-27ab854928f7\") " pod="openshift-marketplace/redhat-operators-dqltg" Oct 04 03:24:33 crc kubenswrapper[4742]: I1004 03:24:33.589438 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/keystone-operator-index-kbkhk" Oct 04 03:24:33 crc kubenswrapper[4742]: I1004 03:24:33.660682 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90332c58-6092-4a1d-b84f-27ab854928f7-catalog-content\") pod \"redhat-operators-dqltg\" (UID: \"90332c58-6092-4a1d-b84f-27ab854928f7\") " pod="openshift-marketplace/redhat-operators-dqltg" Oct 04 03:24:33 crc kubenswrapper[4742]: I1004 03:24:33.660782 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbt2n\" (UniqueName: \"kubernetes.io/projected/90332c58-6092-4a1d-b84f-27ab854928f7-kube-api-access-nbt2n\") pod \"redhat-operators-dqltg\" (UID: \"90332c58-6092-4a1d-b84f-27ab854928f7\") " pod="openshift-marketplace/redhat-operators-dqltg" Oct 04 03:24:33 crc kubenswrapper[4742]: I1004 03:24:33.660818 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90332c58-6092-4a1d-b84f-27ab854928f7-utilities\") pod \"redhat-operators-dqltg\" (UID: \"90332c58-6092-4a1d-b84f-27ab854928f7\") " pod="openshift-marketplace/redhat-operators-dqltg" Oct 04 03:24:33 crc kubenswrapper[4742]: I1004 03:24:33.661420 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90332c58-6092-4a1d-b84f-27ab854928f7-utilities\") pod \"redhat-operators-dqltg\" (UID: \"90332c58-6092-4a1d-b84f-27ab854928f7\") " pod="openshift-marketplace/redhat-operators-dqltg" Oct 04 03:24:33 crc kubenswrapper[4742]: I1004 03:24:33.661425 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90332c58-6092-4a1d-b84f-27ab854928f7-catalog-content\") pod \"redhat-operators-dqltg\" (UID: \"90332c58-6092-4a1d-b84f-27ab854928f7\") " pod="openshift-marketplace/redhat-operators-dqltg" Oct 04 03:24:33 crc kubenswrapper[4742]: I1004 03:24:33.694070 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbt2n\" (UniqueName: \"kubernetes.io/projected/90332c58-6092-4a1d-b84f-27ab854928f7-kube-api-access-nbt2n\") pod \"redhat-operators-dqltg\" (UID: \"90332c58-6092-4a1d-b84f-27ab854928f7\") " pod="openshift-marketplace/redhat-operators-dqltg" Oct 04 03:24:33 crc kubenswrapper[4742]: I1004 03:24:33.736254 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dqltg" Oct 04 03:24:33 crc kubenswrapper[4742]: I1004 03:24:33.897651 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-index-kbkhk" Oct 04 03:24:34 crc kubenswrapper[4742]: I1004 03:24:34.151886 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dqltg"] Oct 04 03:24:34 crc kubenswrapper[4742]: I1004 03:24:34.867954 4742 generic.go:334] "Generic (PLEG): container finished" podID="90332c58-6092-4a1d-b84f-27ab854928f7" containerID="8810eca6ed895027bd36208908fbef747b98c7b9a33fb94368af59345b8f98a1" exitCode=0 Oct 04 03:24:34 crc kubenswrapper[4742]: I1004 03:24:34.884186 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqltg" event={"ID":"90332c58-6092-4a1d-b84f-27ab854928f7","Type":"ContainerDied","Data":"8810eca6ed895027bd36208908fbef747b98c7b9a33fb94368af59345b8f98a1"} Oct 04 03:24:34 crc kubenswrapper[4742]: I1004 03:24:34.884463 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqltg" event={"ID":"90332c58-6092-4a1d-b84f-27ab854928f7","Type":"ContainerStarted","Data":"ae0697f71e03e4130c2843e793b411aed9ce664549b585a9bd31c446e4b215cb"} Oct 04 03:24:35 crc kubenswrapper[4742]: I1004 03:24:35.879982 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqltg" event={"ID":"90332c58-6092-4a1d-b84f-27ab854928f7","Type":"ContainerStarted","Data":"c467dbefdcc96fbdf8a63fe00b78942483d98e6934eabd2489ddf787b5905111"} Oct 04 03:24:36 crc kubenswrapper[4742]: I1004 03:24:36.141613 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-58ppm" Oct 04 03:24:36 crc kubenswrapper[4742]: I1004 03:24:36.141691 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-58ppm" Oct 04 03:24:36 crc kubenswrapper[4742]: I1004 03:24:36.225881 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-58ppm" Oct 04 03:24:36 crc kubenswrapper[4742]: I1004 03:24:36.892879 4742 generic.go:334] "Generic (PLEG): container finished" podID="90332c58-6092-4a1d-b84f-27ab854928f7" containerID="c467dbefdcc96fbdf8a63fe00b78942483d98e6934eabd2489ddf787b5905111" exitCode=0 Oct 04 03:24:36 crc kubenswrapper[4742]: I1004 03:24:36.893105 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqltg" event={"ID":"90332c58-6092-4a1d-b84f-27ab854928f7","Type":"ContainerDied","Data":"c467dbefdcc96fbdf8a63fe00b78942483d98e6934eabd2489ddf787b5905111"} Oct 04 03:24:37 crc kubenswrapper[4742]: I1004 03:24:37.897853 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8"] Oct 04 03:24:37 crc kubenswrapper[4742]: I1004 03:24:37.904946 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8"] Oct 04 03:24:37 crc kubenswrapper[4742]: I1004 03:24:37.905123 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8" Oct 04 03:24:37 crc kubenswrapper[4742]: I1004 03:24:37.908923 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-625gw" Oct 04 03:24:37 crc kubenswrapper[4742]: I1004 03:24:37.930346 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqltg" event={"ID":"90332c58-6092-4a1d-b84f-27ab854928f7","Type":"ContainerStarted","Data":"4b611ec0d5f43a41a0231a3e4ffb84ca37576284511c4101df88f028f9b3fb27"} Oct 04 03:24:37 crc kubenswrapper[4742]: I1004 03:24:37.957482 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dqltg" podStartSLOduration=2.410766369 podStartE2EDuration="4.957454939s" podCreationTimestamp="2025-10-04 03:24:33 +0000 UTC" firstStartedPulling="2025-10-04 03:24:34.870080853 +0000 UTC m=+840.802418029" lastFinishedPulling="2025-10-04 03:24:37.416769383 +0000 UTC m=+843.349106599" observedRunningTime="2025-10-04 03:24:37.952642305 +0000 UTC m=+843.884979521" watchObservedRunningTime="2025-10-04 03:24:37.957454939 +0000 UTC m=+843.889792115" Oct 04 03:24:38 crc kubenswrapper[4742]: I1004 03:24:38.060061 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgnsj\" (UniqueName: \"kubernetes.io/projected/639d681e-3475-410c-b738-6f3be2be5b4b-kube-api-access-dgnsj\") pod \"d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8\" (UID: \"639d681e-3475-410c-b738-6f3be2be5b4b\") " pod="openstack-operators/d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8" Oct 04 03:24:38 crc kubenswrapper[4742]: I1004 03:24:38.060119 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/639d681e-3475-410c-b738-6f3be2be5b4b-bundle\") pod \"d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8\" (UID: \"639d681e-3475-410c-b738-6f3be2be5b4b\") " pod="openstack-operators/d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8" Oct 04 03:24:38 crc kubenswrapper[4742]: I1004 03:24:38.060211 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/639d681e-3475-410c-b738-6f3be2be5b4b-util\") pod \"d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8\" (UID: \"639d681e-3475-410c-b738-6f3be2be5b4b\") " pod="openstack-operators/d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8" Oct 04 03:24:38 crc kubenswrapper[4742]: I1004 03:24:38.161578 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/639d681e-3475-410c-b738-6f3be2be5b4b-util\") pod \"d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8\" (UID: \"639d681e-3475-410c-b738-6f3be2be5b4b\") " pod="openstack-operators/d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8" Oct 04 03:24:38 crc kubenswrapper[4742]: I1004 03:24:38.161724 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgnsj\" (UniqueName: \"kubernetes.io/projected/639d681e-3475-410c-b738-6f3be2be5b4b-kube-api-access-dgnsj\") pod \"d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8\" (UID: \"639d681e-3475-410c-b738-6f3be2be5b4b\") " pod="openstack-operators/d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8" Oct 04 03:24:38 crc kubenswrapper[4742]: I1004 03:24:38.161761 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/639d681e-3475-410c-b738-6f3be2be5b4b-bundle\") pod \"d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8\" (UID: \"639d681e-3475-410c-b738-6f3be2be5b4b\") " pod="openstack-operators/d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8" Oct 04 03:24:38 crc kubenswrapper[4742]: I1004 03:24:38.162378 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/639d681e-3475-410c-b738-6f3be2be5b4b-bundle\") pod \"d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8\" (UID: \"639d681e-3475-410c-b738-6f3be2be5b4b\") " pod="openstack-operators/d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8" Oct 04 03:24:38 crc kubenswrapper[4742]: I1004 03:24:38.162441 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/639d681e-3475-410c-b738-6f3be2be5b4b-util\") pod \"d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8\" (UID: \"639d681e-3475-410c-b738-6f3be2be5b4b\") " pod="openstack-operators/d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8" Oct 04 03:24:38 crc kubenswrapper[4742]: I1004 03:24:38.186981 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgnsj\" (UniqueName: \"kubernetes.io/projected/639d681e-3475-410c-b738-6f3be2be5b4b-kube-api-access-dgnsj\") pod \"d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8\" (UID: \"639d681e-3475-410c-b738-6f3be2be5b4b\") " pod="openstack-operators/d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8" Oct 04 03:24:38 crc kubenswrapper[4742]: I1004 03:24:38.224919 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8" Oct 04 03:24:38 crc kubenswrapper[4742]: I1004 03:24:38.727113 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8"] Oct 04 03:24:38 crc kubenswrapper[4742]: W1004 03:24:38.742570 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod639d681e_3475_410c_b738_6f3be2be5b4b.slice/crio-a556dc0c338a91fb7644ec010b73429f47fc9cd811a34a3ead92996a126ee424 WatchSource:0}: Error finding container a556dc0c338a91fb7644ec010b73429f47fc9cd811a34a3ead92996a126ee424: Status 404 returned error can't find the container with id a556dc0c338a91fb7644ec010b73429f47fc9cd811a34a3ead92996a126ee424 Oct 04 03:24:38 crc kubenswrapper[4742]: I1004 03:24:38.944751 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8" event={"ID":"639d681e-3475-410c-b738-6f3be2be5b4b","Type":"ContainerStarted","Data":"c9f91dbd2403f00ef53d3fc612b2a5f336d49cf798dd18d582c6e92d7faaada2"} Oct 04 03:24:38 crc kubenswrapper[4742]: I1004 03:24:38.945119 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8" event={"ID":"639d681e-3475-410c-b738-6f3be2be5b4b","Type":"ContainerStarted","Data":"a556dc0c338a91fb7644ec010b73429f47fc9cd811a34a3ead92996a126ee424"} Oct 04 03:24:39 crc kubenswrapper[4742]: I1004 03:24:39.957557 4742 generic.go:334] "Generic (PLEG): container finished" podID="639d681e-3475-410c-b738-6f3be2be5b4b" containerID="c9f91dbd2403f00ef53d3fc612b2a5f336d49cf798dd18d582c6e92d7faaada2" exitCode=0 Oct 04 03:24:39 crc kubenswrapper[4742]: I1004 03:24:39.957627 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8" event={"ID":"639d681e-3475-410c-b738-6f3be2be5b4b","Type":"ContainerDied","Data":"c9f91dbd2403f00ef53d3fc612b2a5f336d49cf798dd18d582c6e92d7faaada2"} Oct 04 03:24:40 crc kubenswrapper[4742]: I1004 03:24:40.967177 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8" event={"ID":"639d681e-3475-410c-b738-6f3be2be5b4b","Type":"ContainerStarted","Data":"ba310c8740700c8acca9cef5c0331026033a25b38a22a2b5fd6c5053cfc9e8bd"} Oct 04 03:24:41 crc kubenswrapper[4742]: I1004 03:24:41.975574 4742 generic.go:334] "Generic (PLEG): container finished" podID="639d681e-3475-410c-b738-6f3be2be5b4b" containerID="ba310c8740700c8acca9cef5c0331026033a25b38a22a2b5fd6c5053cfc9e8bd" exitCode=0 Oct 04 03:24:41 crc kubenswrapper[4742]: I1004 03:24:41.975638 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8" event={"ID":"639d681e-3475-410c-b738-6f3be2be5b4b","Type":"ContainerDied","Data":"ba310c8740700c8acca9cef5c0331026033a25b38a22a2b5fd6c5053cfc9e8bd"} Oct 04 03:24:42 crc kubenswrapper[4742]: I1004 03:24:42.988494 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8" event={"ID":"639d681e-3475-410c-b738-6f3be2be5b4b","Type":"ContainerStarted","Data":"d8e671653b0001611ab14fcb313e55d83b351c32b431f83d51c7bb9074efde41"} Oct 04 03:24:43 crc kubenswrapper[4742]: I1004 03:24:43.020144 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8" podStartSLOduration=5.196014905 podStartE2EDuration="6.020109349s" podCreationTimestamp="2025-10-04 03:24:37 +0000 UTC" firstStartedPulling="2025-10-04 03:24:39.959687587 +0000 UTC m=+845.892024803" lastFinishedPulling="2025-10-04 03:24:40.783782051 +0000 UTC m=+846.716119247" observedRunningTime="2025-10-04 03:24:43.013437628 +0000 UTC m=+848.945774824" watchObservedRunningTime="2025-10-04 03:24:43.020109349 +0000 UTC m=+848.952446535" Oct 04 03:24:43 crc kubenswrapper[4742]: I1004 03:24:43.736787 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dqltg" Oct 04 03:24:43 crc kubenswrapper[4742]: I1004 03:24:43.737215 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dqltg" Oct 04 03:24:43 crc kubenswrapper[4742]: I1004 03:24:43.814337 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dqltg" Oct 04 03:24:44 crc kubenswrapper[4742]: I1004 03:24:44.013314 4742 generic.go:334] "Generic (PLEG): container finished" podID="639d681e-3475-410c-b738-6f3be2be5b4b" containerID="d8e671653b0001611ab14fcb313e55d83b351c32b431f83d51c7bb9074efde41" exitCode=0 Oct 04 03:24:44 crc kubenswrapper[4742]: I1004 03:24:44.013427 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8" event={"ID":"639d681e-3475-410c-b738-6f3be2be5b4b","Type":"ContainerDied","Data":"d8e671653b0001611ab14fcb313e55d83b351c32b431f83d51c7bb9074efde41"} Oct 04 03:24:44 crc kubenswrapper[4742]: I1004 03:24:44.092092 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dqltg" Oct 04 03:24:44 crc kubenswrapper[4742]: I1004 03:24:44.872775 4742 patch_prober.go:28] interesting pod/machine-config-daemon-qdt7v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:24:44 crc kubenswrapper[4742]: I1004 03:24:44.872879 4742 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:24:44 crc kubenswrapper[4742]: I1004 03:24:44.872969 4742 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" Oct 04 03:24:44 crc kubenswrapper[4742]: I1004 03:24:44.874468 4742 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ffcec0a1f833323e5678d3dde9b3cc1a6ca05ee38cfe04eae68bdf431a24a5f2"} pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 03:24:44 crc kubenswrapper[4742]: I1004 03:24:44.874656 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerName="machine-config-daemon" containerID="cri-o://ffcec0a1f833323e5678d3dde9b3cc1a6ca05ee38cfe04eae68bdf431a24a5f2" gracePeriod=600 Oct 04 03:24:45 crc kubenswrapper[4742]: I1004 03:24:45.444208 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8" Oct 04 03:24:45 crc kubenswrapper[4742]: I1004 03:24:45.592261 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgnsj\" (UniqueName: \"kubernetes.io/projected/639d681e-3475-410c-b738-6f3be2be5b4b-kube-api-access-dgnsj\") pod \"639d681e-3475-410c-b738-6f3be2be5b4b\" (UID: \"639d681e-3475-410c-b738-6f3be2be5b4b\") " Oct 04 03:24:45 crc kubenswrapper[4742]: I1004 03:24:45.594011 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/639d681e-3475-410c-b738-6f3be2be5b4b-util\") pod \"639d681e-3475-410c-b738-6f3be2be5b4b\" (UID: \"639d681e-3475-410c-b738-6f3be2be5b4b\") " Oct 04 03:24:45 crc kubenswrapper[4742]: I1004 03:24:45.594056 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/639d681e-3475-410c-b738-6f3be2be5b4b-bundle\") pod \"639d681e-3475-410c-b738-6f3be2be5b4b\" (UID: \"639d681e-3475-410c-b738-6f3be2be5b4b\") " Oct 04 03:24:45 crc kubenswrapper[4742]: I1004 03:24:45.595768 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/639d681e-3475-410c-b738-6f3be2be5b4b-bundle" (OuterVolumeSpecName: "bundle") pod "639d681e-3475-410c-b738-6f3be2be5b4b" (UID: "639d681e-3475-410c-b738-6f3be2be5b4b"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:24:45 crc kubenswrapper[4742]: I1004 03:24:45.604397 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/639d681e-3475-410c-b738-6f3be2be5b4b-util" (OuterVolumeSpecName: "util") pod "639d681e-3475-410c-b738-6f3be2be5b4b" (UID: "639d681e-3475-410c-b738-6f3be2be5b4b"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:24:45 crc kubenswrapper[4742]: I1004 03:24:45.605111 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/639d681e-3475-410c-b738-6f3be2be5b4b-kube-api-access-dgnsj" (OuterVolumeSpecName: "kube-api-access-dgnsj") pod "639d681e-3475-410c-b738-6f3be2be5b4b" (UID: "639d681e-3475-410c-b738-6f3be2be5b4b"). InnerVolumeSpecName "kube-api-access-dgnsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:24:45 crc kubenswrapper[4742]: I1004 03:24:45.696601 4742 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/639d681e-3475-410c-b738-6f3be2be5b4b-util\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:45 crc kubenswrapper[4742]: I1004 03:24:45.696887 4742 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/639d681e-3475-410c-b738-6f3be2be5b4b-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:45 crc kubenswrapper[4742]: I1004 03:24:45.696958 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgnsj\" (UniqueName: \"kubernetes.io/projected/639d681e-3475-410c-b738-6f3be2be5b4b-kube-api-access-dgnsj\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:45 crc kubenswrapper[4742]: I1004 03:24:45.796401 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dqltg"] Oct 04 03:24:46 crc kubenswrapper[4742]: I1004 03:24:46.034257 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8" event={"ID":"639d681e-3475-410c-b738-6f3be2be5b4b","Type":"ContainerDied","Data":"a556dc0c338a91fb7644ec010b73429f47fc9cd811a34a3ead92996a126ee424"} Oct 04 03:24:46 crc kubenswrapper[4742]: I1004 03:24:46.034984 4742 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a556dc0c338a91fb7644ec010b73429f47fc9cd811a34a3ead92996a126ee424" Oct 04 03:24:46 crc kubenswrapper[4742]: I1004 03:24:46.034358 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8" Oct 04 03:24:46 crc kubenswrapper[4742]: I1004 03:24:46.038429 4742 generic.go:334] "Generic (PLEG): container finished" podID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerID="ffcec0a1f833323e5678d3dde9b3cc1a6ca05ee38cfe04eae68bdf431a24a5f2" exitCode=0 Oct 04 03:24:46 crc kubenswrapper[4742]: I1004 03:24:46.038518 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" event={"ID":"989ea247-7ac9-4ade-8dcb-2632fef4f83e","Type":"ContainerDied","Data":"ffcec0a1f833323e5678d3dde9b3cc1a6ca05ee38cfe04eae68bdf431a24a5f2"} Oct 04 03:24:46 crc kubenswrapper[4742]: I1004 03:24:46.038594 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" event={"ID":"989ea247-7ac9-4ade-8dcb-2632fef4f83e","Type":"ContainerStarted","Data":"85a4c342170ba17dd23c083b2ba6f556cfcde17e4bf847c7bc6921140b4883ad"} Oct 04 03:24:46 crc kubenswrapper[4742]: I1004 03:24:46.038648 4742 scope.go:117] "RemoveContainer" containerID="716e86aaee18fb777354d280f34d7ae9aea9909aae176a0c4c726413f1e1cb06" Oct 04 03:24:46 crc kubenswrapper[4742]: I1004 03:24:46.187812 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-58ppm" Oct 04 03:24:47 crc kubenswrapper[4742]: I1004 03:24:47.049177 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dqltg" podUID="90332c58-6092-4a1d-b84f-27ab854928f7" containerName="registry-server" containerID="cri-o://4b611ec0d5f43a41a0231a3e4ffb84ca37576284511c4101df88f028f9b3fb27" gracePeriod=2 Oct 04 03:24:47 crc kubenswrapper[4742]: I1004 03:24:47.525095 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dqltg" Oct 04 03:24:47 crc kubenswrapper[4742]: I1004 03:24:47.594595 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-58ppm"] Oct 04 03:24:47 crc kubenswrapper[4742]: I1004 03:24:47.594913 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-58ppm" podUID="125f69b5-d78b-4241-b3d6-c1a151d567e0" containerName="registry-server" containerID="cri-o://96fc5bf80a98cd07699381311b67bdd7694e7ce11da0c6786c54647ae556e981" gracePeriod=2 Oct 04 03:24:47 crc kubenswrapper[4742]: I1004 03:24:47.629051 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90332c58-6092-4a1d-b84f-27ab854928f7-utilities\") pod \"90332c58-6092-4a1d-b84f-27ab854928f7\" (UID: \"90332c58-6092-4a1d-b84f-27ab854928f7\") " Oct 04 03:24:47 crc kubenswrapper[4742]: I1004 03:24:47.629171 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90332c58-6092-4a1d-b84f-27ab854928f7-catalog-content\") pod \"90332c58-6092-4a1d-b84f-27ab854928f7\" (UID: \"90332c58-6092-4a1d-b84f-27ab854928f7\") " Oct 04 03:24:47 crc kubenswrapper[4742]: I1004 03:24:47.629264 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbt2n\" (UniqueName: \"kubernetes.io/projected/90332c58-6092-4a1d-b84f-27ab854928f7-kube-api-access-nbt2n\") pod \"90332c58-6092-4a1d-b84f-27ab854928f7\" (UID: \"90332c58-6092-4a1d-b84f-27ab854928f7\") " Oct 04 03:24:47 crc kubenswrapper[4742]: I1004 03:24:47.630023 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90332c58-6092-4a1d-b84f-27ab854928f7-utilities" (OuterVolumeSpecName: "utilities") pod "90332c58-6092-4a1d-b84f-27ab854928f7" (UID: "90332c58-6092-4a1d-b84f-27ab854928f7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:24:47 crc kubenswrapper[4742]: I1004 03:24:47.637311 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90332c58-6092-4a1d-b84f-27ab854928f7-kube-api-access-nbt2n" (OuterVolumeSpecName: "kube-api-access-nbt2n") pod "90332c58-6092-4a1d-b84f-27ab854928f7" (UID: "90332c58-6092-4a1d-b84f-27ab854928f7"). InnerVolumeSpecName "kube-api-access-nbt2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:24:47 crc kubenswrapper[4742]: I1004 03:24:47.723788 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90332c58-6092-4a1d-b84f-27ab854928f7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "90332c58-6092-4a1d-b84f-27ab854928f7" (UID: "90332c58-6092-4a1d-b84f-27ab854928f7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:24:47 crc kubenswrapper[4742]: I1004 03:24:47.731182 4742 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90332c58-6092-4a1d-b84f-27ab854928f7-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:47 crc kubenswrapper[4742]: I1004 03:24:47.731215 4742 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90332c58-6092-4a1d-b84f-27ab854928f7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:47 crc kubenswrapper[4742]: I1004 03:24:47.731226 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbt2n\" (UniqueName: \"kubernetes.io/projected/90332c58-6092-4a1d-b84f-27ab854928f7-kube-api-access-nbt2n\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.026903 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-58ppm" Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.065414 4742 generic.go:334] "Generic (PLEG): container finished" podID="125f69b5-d78b-4241-b3d6-c1a151d567e0" containerID="96fc5bf80a98cd07699381311b67bdd7694e7ce11da0c6786c54647ae556e981" exitCode=0 Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.065521 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-58ppm" Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.065536 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-58ppm" event={"ID":"125f69b5-d78b-4241-b3d6-c1a151d567e0","Type":"ContainerDied","Data":"96fc5bf80a98cd07699381311b67bdd7694e7ce11da0c6786c54647ae556e981"} Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.065601 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-58ppm" event={"ID":"125f69b5-d78b-4241-b3d6-c1a151d567e0","Type":"ContainerDied","Data":"a7456aa97f293c6ff84e50a9a3d9ec289835cef7eda10e2b8035a9fa10b64cf7"} Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.065641 4742 scope.go:117] "RemoveContainer" containerID="96fc5bf80a98cd07699381311b67bdd7694e7ce11da0c6786c54647ae556e981" Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.070048 4742 generic.go:334] "Generic (PLEG): container finished" podID="90332c58-6092-4a1d-b84f-27ab854928f7" containerID="4b611ec0d5f43a41a0231a3e4ffb84ca37576284511c4101df88f028f9b3fb27" exitCode=0 Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.070159 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqltg" event={"ID":"90332c58-6092-4a1d-b84f-27ab854928f7","Type":"ContainerDied","Data":"4b611ec0d5f43a41a0231a3e4ffb84ca37576284511c4101df88f028f9b3fb27"} Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.070199 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqltg" event={"ID":"90332c58-6092-4a1d-b84f-27ab854928f7","Type":"ContainerDied","Data":"ae0697f71e03e4130c2843e793b411aed9ce664549b585a9bd31c446e4b215cb"} Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.070217 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dqltg" Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.094529 4742 scope.go:117] "RemoveContainer" containerID="b2de163d5213a3ba66c6b26e9140095000c49edf514825393252e1799927bd56" Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.124971 4742 scope.go:117] "RemoveContainer" containerID="21c3ca01c55e96e1e5315b61c3f48794019088830fcd21656248adfbb05221e5" Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.128656 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dqltg"] Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.135439 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dqltg"] Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.150301 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/125f69b5-d78b-4241-b3d6-c1a151d567e0-utilities\") pod \"125f69b5-d78b-4241-b3d6-c1a151d567e0\" (UID: \"125f69b5-d78b-4241-b3d6-c1a151d567e0\") " Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.150440 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/125f69b5-d78b-4241-b3d6-c1a151d567e0-catalog-content\") pod \"125f69b5-d78b-4241-b3d6-c1a151d567e0\" (UID: \"125f69b5-d78b-4241-b3d6-c1a151d567e0\") " Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.150600 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvb6k\" (UniqueName: \"kubernetes.io/projected/125f69b5-d78b-4241-b3d6-c1a151d567e0-kube-api-access-vvb6k\") pod \"125f69b5-d78b-4241-b3d6-c1a151d567e0\" (UID: \"125f69b5-d78b-4241-b3d6-c1a151d567e0\") " Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.163786 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/125f69b5-d78b-4241-b3d6-c1a151d567e0-utilities" (OuterVolumeSpecName: "utilities") pod "125f69b5-d78b-4241-b3d6-c1a151d567e0" (UID: "125f69b5-d78b-4241-b3d6-c1a151d567e0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.163812 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/125f69b5-d78b-4241-b3d6-c1a151d567e0-kube-api-access-vvb6k" (OuterVolumeSpecName: "kube-api-access-vvb6k") pod "125f69b5-d78b-4241-b3d6-c1a151d567e0" (UID: "125f69b5-d78b-4241-b3d6-c1a151d567e0"). InnerVolumeSpecName "kube-api-access-vvb6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.164093 4742 scope.go:117] "RemoveContainer" containerID="96fc5bf80a98cd07699381311b67bdd7694e7ce11da0c6786c54647ae556e981" Oct 04 03:24:48 crc kubenswrapper[4742]: E1004 03:24:48.164895 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96fc5bf80a98cd07699381311b67bdd7694e7ce11da0c6786c54647ae556e981\": container with ID starting with 96fc5bf80a98cd07699381311b67bdd7694e7ce11da0c6786c54647ae556e981 not found: ID does not exist" containerID="96fc5bf80a98cd07699381311b67bdd7694e7ce11da0c6786c54647ae556e981" Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.164954 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96fc5bf80a98cd07699381311b67bdd7694e7ce11da0c6786c54647ae556e981"} err="failed to get container status \"96fc5bf80a98cd07699381311b67bdd7694e7ce11da0c6786c54647ae556e981\": rpc error: code = NotFound desc = could not find container \"96fc5bf80a98cd07699381311b67bdd7694e7ce11da0c6786c54647ae556e981\": container with ID starting with 96fc5bf80a98cd07699381311b67bdd7694e7ce11da0c6786c54647ae556e981 not found: ID does not exist" Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.164993 4742 scope.go:117] "RemoveContainer" containerID="b2de163d5213a3ba66c6b26e9140095000c49edf514825393252e1799927bd56" Oct 04 03:24:48 crc kubenswrapper[4742]: E1004 03:24:48.165556 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2de163d5213a3ba66c6b26e9140095000c49edf514825393252e1799927bd56\": container with ID starting with b2de163d5213a3ba66c6b26e9140095000c49edf514825393252e1799927bd56 not found: ID does not exist" containerID="b2de163d5213a3ba66c6b26e9140095000c49edf514825393252e1799927bd56" Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.165646 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2de163d5213a3ba66c6b26e9140095000c49edf514825393252e1799927bd56"} err="failed to get container status \"b2de163d5213a3ba66c6b26e9140095000c49edf514825393252e1799927bd56\": rpc error: code = NotFound desc = could not find container \"b2de163d5213a3ba66c6b26e9140095000c49edf514825393252e1799927bd56\": container with ID starting with b2de163d5213a3ba66c6b26e9140095000c49edf514825393252e1799927bd56 not found: ID does not exist" Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.165711 4742 scope.go:117] "RemoveContainer" containerID="21c3ca01c55e96e1e5315b61c3f48794019088830fcd21656248adfbb05221e5" Oct 04 03:24:48 crc kubenswrapper[4742]: E1004 03:24:48.167463 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21c3ca01c55e96e1e5315b61c3f48794019088830fcd21656248adfbb05221e5\": container with ID starting with 21c3ca01c55e96e1e5315b61c3f48794019088830fcd21656248adfbb05221e5 not found: ID does not exist" containerID="21c3ca01c55e96e1e5315b61c3f48794019088830fcd21656248adfbb05221e5" Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.167509 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21c3ca01c55e96e1e5315b61c3f48794019088830fcd21656248adfbb05221e5"} err="failed to get container status \"21c3ca01c55e96e1e5315b61c3f48794019088830fcd21656248adfbb05221e5\": rpc error: code = NotFound desc = could not find container \"21c3ca01c55e96e1e5315b61c3f48794019088830fcd21656248adfbb05221e5\": container with ID starting with 21c3ca01c55e96e1e5315b61c3f48794019088830fcd21656248adfbb05221e5 not found: ID does not exist" Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.167537 4742 scope.go:117] "RemoveContainer" containerID="4b611ec0d5f43a41a0231a3e4ffb84ca37576284511c4101df88f028f9b3fb27" Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.209962 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/125f69b5-d78b-4241-b3d6-c1a151d567e0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "125f69b5-d78b-4241-b3d6-c1a151d567e0" (UID: "125f69b5-d78b-4241-b3d6-c1a151d567e0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.216817 4742 scope.go:117] "RemoveContainer" containerID="c467dbefdcc96fbdf8a63fe00b78942483d98e6934eabd2489ddf787b5905111" Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.254056 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvb6k\" (UniqueName: \"kubernetes.io/projected/125f69b5-d78b-4241-b3d6-c1a151d567e0-kube-api-access-vvb6k\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.254221 4742 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/125f69b5-d78b-4241-b3d6-c1a151d567e0-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.254245 4742 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/125f69b5-d78b-4241-b3d6-c1a151d567e0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.257684 4742 scope.go:117] "RemoveContainer" containerID="8810eca6ed895027bd36208908fbef747b98c7b9a33fb94368af59345b8f98a1" Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.283761 4742 scope.go:117] "RemoveContainer" containerID="4b611ec0d5f43a41a0231a3e4ffb84ca37576284511c4101df88f028f9b3fb27" Oct 04 03:24:48 crc kubenswrapper[4742]: E1004 03:24:48.284974 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b611ec0d5f43a41a0231a3e4ffb84ca37576284511c4101df88f028f9b3fb27\": container with ID starting with 4b611ec0d5f43a41a0231a3e4ffb84ca37576284511c4101df88f028f9b3fb27 not found: ID does not exist" containerID="4b611ec0d5f43a41a0231a3e4ffb84ca37576284511c4101df88f028f9b3fb27" Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.285064 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b611ec0d5f43a41a0231a3e4ffb84ca37576284511c4101df88f028f9b3fb27"} err="failed to get container status \"4b611ec0d5f43a41a0231a3e4ffb84ca37576284511c4101df88f028f9b3fb27\": rpc error: code = NotFound desc = could not find container \"4b611ec0d5f43a41a0231a3e4ffb84ca37576284511c4101df88f028f9b3fb27\": container with ID starting with 4b611ec0d5f43a41a0231a3e4ffb84ca37576284511c4101df88f028f9b3fb27 not found: ID does not exist" Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.285107 4742 scope.go:117] "RemoveContainer" containerID="c467dbefdcc96fbdf8a63fe00b78942483d98e6934eabd2489ddf787b5905111" Oct 04 03:24:48 crc kubenswrapper[4742]: E1004 03:24:48.285834 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c467dbefdcc96fbdf8a63fe00b78942483d98e6934eabd2489ddf787b5905111\": container with ID starting with c467dbefdcc96fbdf8a63fe00b78942483d98e6934eabd2489ddf787b5905111 not found: ID does not exist" containerID="c467dbefdcc96fbdf8a63fe00b78942483d98e6934eabd2489ddf787b5905111" Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.285922 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c467dbefdcc96fbdf8a63fe00b78942483d98e6934eabd2489ddf787b5905111"} err="failed to get container status \"c467dbefdcc96fbdf8a63fe00b78942483d98e6934eabd2489ddf787b5905111\": rpc error: code = NotFound desc = could not find container \"c467dbefdcc96fbdf8a63fe00b78942483d98e6934eabd2489ddf787b5905111\": container with ID starting with c467dbefdcc96fbdf8a63fe00b78942483d98e6934eabd2489ddf787b5905111 not found: ID does not exist" Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.285983 4742 scope.go:117] "RemoveContainer" containerID="8810eca6ed895027bd36208908fbef747b98c7b9a33fb94368af59345b8f98a1" Oct 04 03:24:48 crc kubenswrapper[4742]: E1004 03:24:48.286633 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8810eca6ed895027bd36208908fbef747b98c7b9a33fb94368af59345b8f98a1\": container with ID starting with 8810eca6ed895027bd36208908fbef747b98c7b9a33fb94368af59345b8f98a1 not found: ID does not exist" containerID="8810eca6ed895027bd36208908fbef747b98c7b9a33fb94368af59345b8f98a1" Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.286670 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8810eca6ed895027bd36208908fbef747b98c7b9a33fb94368af59345b8f98a1"} err="failed to get container status \"8810eca6ed895027bd36208908fbef747b98c7b9a33fb94368af59345b8f98a1\": rpc error: code = NotFound desc = could not find container \"8810eca6ed895027bd36208908fbef747b98c7b9a33fb94368af59345b8f98a1\": container with ID starting with 8810eca6ed895027bd36208908fbef747b98c7b9a33fb94368af59345b8f98a1 not found: ID does not exist" Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.402571 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-58ppm"] Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.406881 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-58ppm"] Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.867127 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="125f69b5-d78b-4241-b3d6-c1a151d567e0" path="/var/lib/kubelet/pods/125f69b5-d78b-4241-b3d6-c1a151d567e0/volumes" Oct 04 03:24:48 crc kubenswrapper[4742]: I1004 03:24:48.868771 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90332c58-6092-4a1d-b84f-27ab854928f7" path="/var/lib/kubelet/pods/90332c58-6092-4a1d-b84f-27ab854928f7/volumes" Oct 04 03:24:52 crc kubenswrapper[4742]: I1004 03:24:52.200869 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-g8xs6"] Oct 04 03:24:52 crc kubenswrapper[4742]: E1004 03:24:52.202142 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90332c58-6092-4a1d-b84f-27ab854928f7" containerName="registry-server" Oct 04 03:24:52 crc kubenswrapper[4742]: I1004 03:24:52.202162 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="90332c58-6092-4a1d-b84f-27ab854928f7" containerName="registry-server" Oct 04 03:24:52 crc kubenswrapper[4742]: E1004 03:24:52.202179 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="125f69b5-d78b-4241-b3d6-c1a151d567e0" containerName="extract-content" Oct 04 03:24:52 crc kubenswrapper[4742]: I1004 03:24:52.202187 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="125f69b5-d78b-4241-b3d6-c1a151d567e0" containerName="extract-content" Oct 04 03:24:52 crc kubenswrapper[4742]: E1004 03:24:52.202206 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90332c58-6092-4a1d-b84f-27ab854928f7" containerName="extract-content" Oct 04 03:24:52 crc kubenswrapper[4742]: I1004 03:24:52.202215 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="90332c58-6092-4a1d-b84f-27ab854928f7" containerName="extract-content" Oct 04 03:24:52 crc kubenswrapper[4742]: E1004 03:24:52.202229 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90332c58-6092-4a1d-b84f-27ab854928f7" containerName="extract-utilities" Oct 04 03:24:52 crc kubenswrapper[4742]: I1004 03:24:52.202237 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="90332c58-6092-4a1d-b84f-27ab854928f7" containerName="extract-utilities" Oct 04 03:24:52 crc kubenswrapper[4742]: E1004 03:24:52.202248 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="125f69b5-d78b-4241-b3d6-c1a151d567e0" containerName="registry-server" Oct 04 03:24:52 crc kubenswrapper[4742]: I1004 03:24:52.202258 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="125f69b5-d78b-4241-b3d6-c1a151d567e0" containerName="registry-server" Oct 04 03:24:52 crc kubenswrapper[4742]: E1004 03:24:52.202297 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="639d681e-3475-410c-b738-6f3be2be5b4b" containerName="pull" Oct 04 03:24:52 crc kubenswrapper[4742]: I1004 03:24:52.202305 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="639d681e-3475-410c-b738-6f3be2be5b4b" containerName="pull" Oct 04 03:24:52 crc kubenswrapper[4742]: E1004 03:24:52.202506 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="639d681e-3475-410c-b738-6f3be2be5b4b" containerName="extract" Oct 04 03:24:52 crc kubenswrapper[4742]: I1004 03:24:52.202515 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="639d681e-3475-410c-b738-6f3be2be5b4b" containerName="extract" Oct 04 03:24:52 crc kubenswrapper[4742]: E1004 03:24:52.202528 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="125f69b5-d78b-4241-b3d6-c1a151d567e0" containerName="extract-utilities" Oct 04 03:24:52 crc kubenswrapper[4742]: I1004 03:24:52.202536 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="125f69b5-d78b-4241-b3d6-c1a151d567e0" containerName="extract-utilities" Oct 04 03:24:52 crc kubenswrapper[4742]: E1004 03:24:52.202551 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="639d681e-3475-410c-b738-6f3be2be5b4b" containerName="util" Oct 04 03:24:52 crc kubenswrapper[4742]: I1004 03:24:52.202558 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="639d681e-3475-410c-b738-6f3be2be5b4b" containerName="util" Oct 04 03:24:52 crc kubenswrapper[4742]: I1004 03:24:52.202695 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="125f69b5-d78b-4241-b3d6-c1a151d567e0" containerName="registry-server" Oct 04 03:24:52 crc kubenswrapper[4742]: I1004 03:24:52.202713 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="90332c58-6092-4a1d-b84f-27ab854928f7" containerName="registry-server" Oct 04 03:24:52 crc kubenswrapper[4742]: I1004 03:24:52.202728 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="639d681e-3475-410c-b738-6f3be2be5b4b" containerName="extract" Oct 04 03:24:52 crc kubenswrapper[4742]: I1004 03:24:52.205449 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g8xs6" Oct 04 03:24:52 crc kubenswrapper[4742]: I1004 03:24:52.226513 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g8xs6"] Oct 04 03:24:52 crc kubenswrapper[4742]: I1004 03:24:52.369397 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfj6t\" (UniqueName: \"kubernetes.io/projected/c44b7e8b-6f70-49ff-9be1-857ff2209913-kube-api-access-qfj6t\") pod \"community-operators-g8xs6\" (UID: \"c44b7e8b-6f70-49ff-9be1-857ff2209913\") " pod="openshift-marketplace/community-operators-g8xs6" Oct 04 03:24:52 crc kubenswrapper[4742]: I1004 03:24:52.369468 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c44b7e8b-6f70-49ff-9be1-857ff2209913-catalog-content\") pod \"community-operators-g8xs6\" (UID: \"c44b7e8b-6f70-49ff-9be1-857ff2209913\") " pod="openshift-marketplace/community-operators-g8xs6" Oct 04 03:24:52 crc kubenswrapper[4742]: I1004 03:24:52.369500 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c44b7e8b-6f70-49ff-9be1-857ff2209913-utilities\") pod \"community-operators-g8xs6\" (UID: \"c44b7e8b-6f70-49ff-9be1-857ff2209913\") " pod="openshift-marketplace/community-operators-g8xs6" Oct 04 03:24:52 crc kubenswrapper[4742]: I1004 03:24:52.470547 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfj6t\" (UniqueName: \"kubernetes.io/projected/c44b7e8b-6f70-49ff-9be1-857ff2209913-kube-api-access-qfj6t\") pod \"community-operators-g8xs6\" (UID: \"c44b7e8b-6f70-49ff-9be1-857ff2209913\") " pod="openshift-marketplace/community-operators-g8xs6" Oct 04 03:24:52 crc kubenswrapper[4742]: I1004 03:24:52.470886 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c44b7e8b-6f70-49ff-9be1-857ff2209913-catalog-content\") pod \"community-operators-g8xs6\" (UID: \"c44b7e8b-6f70-49ff-9be1-857ff2209913\") " pod="openshift-marketplace/community-operators-g8xs6" Oct 04 03:24:52 crc kubenswrapper[4742]: I1004 03:24:52.471037 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c44b7e8b-6f70-49ff-9be1-857ff2209913-utilities\") pod \"community-operators-g8xs6\" (UID: \"c44b7e8b-6f70-49ff-9be1-857ff2209913\") " pod="openshift-marketplace/community-operators-g8xs6" Oct 04 03:24:52 crc kubenswrapper[4742]: I1004 03:24:52.471590 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c44b7e8b-6f70-49ff-9be1-857ff2209913-catalog-content\") pod \"community-operators-g8xs6\" (UID: \"c44b7e8b-6f70-49ff-9be1-857ff2209913\") " pod="openshift-marketplace/community-operators-g8xs6" Oct 04 03:24:52 crc kubenswrapper[4742]: I1004 03:24:52.471831 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c44b7e8b-6f70-49ff-9be1-857ff2209913-utilities\") pod \"community-operators-g8xs6\" (UID: \"c44b7e8b-6f70-49ff-9be1-857ff2209913\") " pod="openshift-marketplace/community-operators-g8xs6" Oct 04 03:24:52 crc kubenswrapper[4742]: I1004 03:24:52.510307 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfj6t\" (UniqueName: \"kubernetes.io/projected/c44b7e8b-6f70-49ff-9be1-857ff2209913-kube-api-access-qfj6t\") pod \"community-operators-g8xs6\" (UID: \"c44b7e8b-6f70-49ff-9be1-857ff2209913\") " pod="openshift-marketplace/community-operators-g8xs6" Oct 04 03:24:52 crc kubenswrapper[4742]: I1004 03:24:52.534092 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g8xs6" Oct 04 03:24:52 crc kubenswrapper[4742]: I1004 03:24:52.785534 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g8xs6"] Oct 04 03:24:53 crc kubenswrapper[4742]: I1004 03:24:53.165081 4742 generic.go:334] "Generic (PLEG): container finished" podID="c44b7e8b-6f70-49ff-9be1-857ff2209913" containerID="702850a871b6e344b5926af819a8b44aab11c822d73185f526ab86968ff3039e" exitCode=0 Oct 04 03:24:53 crc kubenswrapper[4742]: I1004 03:24:53.165150 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g8xs6" event={"ID":"c44b7e8b-6f70-49ff-9be1-857ff2209913","Type":"ContainerDied","Data":"702850a871b6e344b5926af819a8b44aab11c822d73185f526ab86968ff3039e"} Oct 04 03:24:53 crc kubenswrapper[4742]: I1004 03:24:53.165189 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g8xs6" event={"ID":"c44b7e8b-6f70-49ff-9be1-857ff2209913","Type":"ContainerStarted","Data":"af75513046bfe79823dc37a7b32797061a9ff13282f3cbaa90fc853b382ca4b1"} Oct 04 03:24:54 crc kubenswrapper[4742]: I1004 03:24:54.173727 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g8xs6" event={"ID":"c44b7e8b-6f70-49ff-9be1-857ff2209913","Type":"ContainerStarted","Data":"2ff9c14ac06b395d1afa2473e2ad4cb6b08d5e24b3e13f1f56dee259e2fd7004"} Oct 04 03:24:55 crc kubenswrapper[4742]: I1004 03:24:55.186146 4742 generic.go:334] "Generic (PLEG): container finished" podID="c44b7e8b-6f70-49ff-9be1-857ff2209913" containerID="2ff9c14ac06b395d1afa2473e2ad4cb6b08d5e24b3e13f1f56dee259e2fd7004" exitCode=0 Oct 04 03:24:55 crc kubenswrapper[4742]: I1004 03:24:55.186201 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g8xs6" event={"ID":"c44b7e8b-6f70-49ff-9be1-857ff2209913","Type":"ContainerDied","Data":"2ff9c14ac06b395d1afa2473e2ad4cb6b08d5e24b3e13f1f56dee259e2fd7004"} Oct 04 03:24:56 crc kubenswrapper[4742]: I1004 03:24:56.202477 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g8xs6" event={"ID":"c44b7e8b-6f70-49ff-9be1-857ff2209913","Type":"ContainerStarted","Data":"e4aaf63681482a7e5e1ce3b8844356e7742ab9fad3b277e20e7f4d9ce0407219"} Oct 04 03:24:56 crc kubenswrapper[4742]: I1004 03:24:56.227233 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-g8xs6" podStartSLOduration=1.811410504 podStartE2EDuration="4.227208153s" podCreationTimestamp="2025-10-04 03:24:52 +0000 UTC" firstStartedPulling="2025-10-04 03:24:53.170569948 +0000 UTC m=+859.102907124" lastFinishedPulling="2025-10-04 03:24:55.586367597 +0000 UTC m=+861.518704773" observedRunningTime="2025-10-04 03:24:56.224545564 +0000 UTC m=+862.156882760" watchObservedRunningTime="2025-10-04 03:24:56.227208153 +0000 UTC m=+862.159545329" Oct 04 03:24:56 crc kubenswrapper[4742]: I1004 03:24:56.797992 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-9d6cf7d55-znns7"] Oct 04 03:24:56 crc kubenswrapper[4742]: I1004 03:24:56.799514 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-9d6cf7d55-znns7" Oct 04 03:24:56 crc kubenswrapper[4742]: I1004 03:24:56.801772 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-service-cert" Oct 04 03:24:56 crc kubenswrapper[4742]: I1004 03:24:56.802099 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-xkkr5" Oct 04 03:24:56 crc kubenswrapper[4742]: I1004 03:24:56.816471 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-9d6cf7d55-znns7"] Oct 04 03:24:56 crc kubenswrapper[4742]: I1004 03:24:56.845796 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/07343d4b-0199-4669-9e2b-9eb697e8b387-apiservice-cert\") pod \"keystone-operator-controller-manager-9d6cf7d55-znns7\" (UID: \"07343d4b-0199-4669-9e2b-9eb697e8b387\") " pod="openstack-operators/keystone-operator-controller-manager-9d6cf7d55-znns7" Oct 04 03:24:56 crc kubenswrapper[4742]: I1004 03:24:56.845850 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcz8g\" (UniqueName: \"kubernetes.io/projected/07343d4b-0199-4669-9e2b-9eb697e8b387-kube-api-access-pcz8g\") pod \"keystone-operator-controller-manager-9d6cf7d55-znns7\" (UID: \"07343d4b-0199-4669-9e2b-9eb697e8b387\") " pod="openstack-operators/keystone-operator-controller-manager-9d6cf7d55-znns7" Oct 04 03:24:56 crc kubenswrapper[4742]: I1004 03:24:56.845887 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/07343d4b-0199-4669-9e2b-9eb697e8b387-webhook-cert\") pod \"keystone-operator-controller-manager-9d6cf7d55-znns7\" (UID: \"07343d4b-0199-4669-9e2b-9eb697e8b387\") " pod="openstack-operators/keystone-operator-controller-manager-9d6cf7d55-znns7" Oct 04 03:24:56 crc kubenswrapper[4742]: I1004 03:24:56.947766 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/07343d4b-0199-4669-9e2b-9eb697e8b387-apiservice-cert\") pod \"keystone-operator-controller-manager-9d6cf7d55-znns7\" (UID: \"07343d4b-0199-4669-9e2b-9eb697e8b387\") " pod="openstack-operators/keystone-operator-controller-manager-9d6cf7d55-znns7" Oct 04 03:24:56 crc kubenswrapper[4742]: I1004 03:24:56.947854 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcz8g\" (UniqueName: \"kubernetes.io/projected/07343d4b-0199-4669-9e2b-9eb697e8b387-kube-api-access-pcz8g\") pod \"keystone-operator-controller-manager-9d6cf7d55-znns7\" (UID: \"07343d4b-0199-4669-9e2b-9eb697e8b387\") " pod="openstack-operators/keystone-operator-controller-manager-9d6cf7d55-znns7" Oct 04 03:24:56 crc kubenswrapper[4742]: I1004 03:24:56.947905 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/07343d4b-0199-4669-9e2b-9eb697e8b387-webhook-cert\") pod \"keystone-operator-controller-manager-9d6cf7d55-znns7\" (UID: \"07343d4b-0199-4669-9e2b-9eb697e8b387\") " pod="openstack-operators/keystone-operator-controller-manager-9d6cf7d55-znns7" Oct 04 03:24:56 crc kubenswrapper[4742]: I1004 03:24:56.957982 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/07343d4b-0199-4669-9e2b-9eb697e8b387-apiservice-cert\") pod \"keystone-operator-controller-manager-9d6cf7d55-znns7\" (UID: \"07343d4b-0199-4669-9e2b-9eb697e8b387\") " pod="openstack-operators/keystone-operator-controller-manager-9d6cf7d55-znns7" Oct 04 03:24:56 crc kubenswrapper[4742]: I1004 03:24:56.968698 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcz8g\" (UniqueName: \"kubernetes.io/projected/07343d4b-0199-4669-9e2b-9eb697e8b387-kube-api-access-pcz8g\") pod \"keystone-operator-controller-manager-9d6cf7d55-znns7\" (UID: \"07343d4b-0199-4669-9e2b-9eb697e8b387\") " pod="openstack-operators/keystone-operator-controller-manager-9d6cf7d55-znns7" Oct 04 03:24:56 crc kubenswrapper[4742]: I1004 03:24:56.970129 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/07343d4b-0199-4669-9e2b-9eb697e8b387-webhook-cert\") pod \"keystone-operator-controller-manager-9d6cf7d55-znns7\" (UID: \"07343d4b-0199-4669-9e2b-9eb697e8b387\") " pod="openstack-operators/keystone-operator-controller-manager-9d6cf7d55-znns7" Oct 04 03:24:57 crc kubenswrapper[4742]: I1004 03:24:57.125544 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-9d6cf7d55-znns7" Oct 04 03:24:57 crc kubenswrapper[4742]: W1004 03:24:57.637170 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07343d4b_0199_4669_9e2b_9eb697e8b387.slice/crio-d0c2732344b10ef11e51f9c71449c05cf68440979c84682b04e13c42ac0c9416 WatchSource:0}: Error finding container d0c2732344b10ef11e51f9c71449c05cf68440979c84682b04e13c42ac0c9416: Status 404 returned error can't find the container with id d0c2732344b10ef11e51f9c71449c05cf68440979c84682b04e13c42ac0c9416 Oct 04 03:24:57 crc kubenswrapper[4742]: I1004 03:24:57.637843 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-9d6cf7d55-znns7"] Oct 04 03:24:58 crc kubenswrapper[4742]: I1004 03:24:58.218055 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-9d6cf7d55-znns7" event={"ID":"07343d4b-0199-4669-9e2b-9eb697e8b387","Type":"ContainerStarted","Data":"d0c2732344b10ef11e51f9c71449c05cf68440979c84682b04e13c42ac0c9416"} Oct 04 03:25:01 crc kubenswrapper[4742]: I1004 03:25:01.247957 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-9d6cf7d55-znns7" event={"ID":"07343d4b-0199-4669-9e2b-9eb697e8b387","Type":"ContainerStarted","Data":"54534e838760ec4df74bb39582be08adac28e00fce4712f2521c489908bb1914"} Oct 04 03:25:01 crc kubenswrapper[4742]: I1004 03:25:01.248933 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-9d6cf7d55-znns7" event={"ID":"07343d4b-0199-4669-9e2b-9eb697e8b387","Type":"ContainerStarted","Data":"52d4b6f81700b8df4dcfb52d072aef207821d31d628cfb918f205673501a1cb8"} Oct 04 03:25:01 crc kubenswrapper[4742]: I1004 03:25:01.248970 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-9d6cf7d55-znns7" Oct 04 03:25:02 crc kubenswrapper[4742]: I1004 03:25:02.260649 4742 generic.go:334] "Generic (PLEG): container finished" podID="8e6c595f-148d-419e-8a72-a991c62df809" containerID="de3e9b1fa72f15a5970dc89363224774e34de58289ec158a7add75fd811b8d23" exitCode=0 Oct 04 03:25:02 crc kubenswrapper[4742]: I1004 03:25:02.260756 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"8e6c595f-148d-419e-8a72-a991c62df809","Type":"ContainerDied","Data":"de3e9b1fa72f15a5970dc89363224774e34de58289ec158a7add75fd811b8d23"} Oct 04 03:25:02 crc kubenswrapper[4742]: I1004 03:25:02.306191 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-9d6cf7d55-znns7" podStartSLOduration=3.064512178 podStartE2EDuration="6.306154248s" podCreationTimestamp="2025-10-04 03:24:56 +0000 UTC" firstStartedPulling="2025-10-04 03:24:57.639117797 +0000 UTC m=+863.571454973" lastFinishedPulling="2025-10-04 03:25:00.880759867 +0000 UTC m=+866.813097043" observedRunningTime="2025-10-04 03:25:01.280193205 +0000 UTC m=+867.212530401" watchObservedRunningTime="2025-10-04 03:25:02.306154248 +0000 UTC m=+868.238491464" Oct 04 03:25:02 crc kubenswrapper[4742]: I1004 03:25:02.534993 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-g8xs6" Oct 04 03:25:02 crc kubenswrapper[4742]: I1004 03:25:02.535603 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-g8xs6" Oct 04 03:25:02 crc kubenswrapper[4742]: I1004 03:25:02.600436 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-g8xs6" Oct 04 03:25:03 crc kubenswrapper[4742]: I1004 03:25:03.272545 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"8e6c595f-148d-419e-8a72-a991c62df809","Type":"ContainerStarted","Data":"b4abb766b03ce81e6fcb2a07b51c4fe0fd95b17be9af6398031fe168958fe243"} Oct 04 03:25:03 crc kubenswrapper[4742]: I1004 03:25:03.273570 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 03:25:03 crc kubenswrapper[4742]: I1004 03:25:03.308365 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/rabbitmq-server-0" podStartSLOduration=36.650232358 podStartE2EDuration="43.30833634s" podCreationTimestamp="2025-10-04 03:24:20 +0000 UTC" firstStartedPulling="2025-10-04 03:24:21.806585218 +0000 UTC m=+827.738922394" lastFinishedPulling="2025-10-04 03:24:28.46468919 +0000 UTC m=+834.397026376" observedRunningTime="2025-10-04 03:25:03.305491206 +0000 UTC m=+869.237828412" watchObservedRunningTime="2025-10-04 03:25:03.30833634 +0000 UTC m=+869.240673526" Oct 04 03:25:03 crc kubenswrapper[4742]: I1004 03:25:03.359241 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-g8xs6" Oct 04 03:25:06 crc kubenswrapper[4742]: I1004 03:25:06.993704 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g8xs6"] Oct 04 03:25:06 crc kubenswrapper[4742]: I1004 03:25:06.994553 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-g8xs6" podUID="c44b7e8b-6f70-49ff-9be1-857ff2209913" containerName="registry-server" containerID="cri-o://e4aaf63681482a7e5e1ce3b8844356e7742ab9fad3b277e20e7f4d9ce0407219" gracePeriod=2 Oct 04 03:25:07 crc kubenswrapper[4742]: I1004 03:25:07.136411 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-9d6cf7d55-znns7" Oct 04 03:25:07 crc kubenswrapper[4742]: I1004 03:25:07.307790 4742 generic.go:334] "Generic (PLEG): container finished" podID="c44b7e8b-6f70-49ff-9be1-857ff2209913" containerID="e4aaf63681482a7e5e1ce3b8844356e7742ab9fad3b277e20e7f4d9ce0407219" exitCode=0 Oct 04 03:25:07 crc kubenswrapper[4742]: I1004 03:25:07.307849 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g8xs6" event={"ID":"c44b7e8b-6f70-49ff-9be1-857ff2209913","Type":"ContainerDied","Data":"e4aaf63681482a7e5e1ce3b8844356e7742ab9fad3b277e20e7f4d9ce0407219"} Oct 04 03:25:07 crc kubenswrapper[4742]: I1004 03:25:07.491104 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g8xs6" Oct 04 03:25:07 crc kubenswrapper[4742]: I1004 03:25:07.616072 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfj6t\" (UniqueName: \"kubernetes.io/projected/c44b7e8b-6f70-49ff-9be1-857ff2209913-kube-api-access-qfj6t\") pod \"c44b7e8b-6f70-49ff-9be1-857ff2209913\" (UID: \"c44b7e8b-6f70-49ff-9be1-857ff2209913\") " Oct 04 03:25:07 crc kubenswrapper[4742]: I1004 03:25:07.616160 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c44b7e8b-6f70-49ff-9be1-857ff2209913-catalog-content\") pod \"c44b7e8b-6f70-49ff-9be1-857ff2209913\" (UID: \"c44b7e8b-6f70-49ff-9be1-857ff2209913\") " Oct 04 03:25:07 crc kubenswrapper[4742]: I1004 03:25:07.616235 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c44b7e8b-6f70-49ff-9be1-857ff2209913-utilities\") pod \"c44b7e8b-6f70-49ff-9be1-857ff2209913\" (UID: \"c44b7e8b-6f70-49ff-9be1-857ff2209913\") " Oct 04 03:25:07 crc kubenswrapper[4742]: I1004 03:25:07.617528 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c44b7e8b-6f70-49ff-9be1-857ff2209913-utilities" (OuterVolumeSpecName: "utilities") pod "c44b7e8b-6f70-49ff-9be1-857ff2209913" (UID: "c44b7e8b-6f70-49ff-9be1-857ff2209913"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:25:07 crc kubenswrapper[4742]: I1004 03:25:07.624218 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c44b7e8b-6f70-49ff-9be1-857ff2209913-kube-api-access-qfj6t" (OuterVolumeSpecName: "kube-api-access-qfj6t") pod "c44b7e8b-6f70-49ff-9be1-857ff2209913" (UID: "c44b7e8b-6f70-49ff-9be1-857ff2209913"). InnerVolumeSpecName "kube-api-access-qfj6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:07 crc kubenswrapper[4742]: I1004 03:25:07.679612 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c44b7e8b-6f70-49ff-9be1-857ff2209913-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c44b7e8b-6f70-49ff-9be1-857ff2209913" (UID: "c44b7e8b-6f70-49ff-9be1-857ff2209913"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:25:07 crc kubenswrapper[4742]: I1004 03:25:07.717426 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfj6t\" (UniqueName: \"kubernetes.io/projected/c44b7e8b-6f70-49ff-9be1-857ff2209913-kube-api-access-qfj6t\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:07 crc kubenswrapper[4742]: I1004 03:25:07.717472 4742 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c44b7e8b-6f70-49ff-9be1-857ff2209913-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:07 crc kubenswrapper[4742]: I1004 03:25:07.717482 4742 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c44b7e8b-6f70-49ff-9be1-857ff2209913-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:08 crc kubenswrapper[4742]: I1004 03:25:08.320000 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g8xs6" event={"ID":"c44b7e8b-6f70-49ff-9be1-857ff2209913","Type":"ContainerDied","Data":"af75513046bfe79823dc37a7b32797061a9ff13282f3cbaa90fc853b382ca4b1"} Oct 04 03:25:08 crc kubenswrapper[4742]: I1004 03:25:08.320069 4742 scope.go:117] "RemoveContainer" containerID="e4aaf63681482a7e5e1ce3b8844356e7742ab9fad3b277e20e7f4d9ce0407219" Oct 04 03:25:08 crc kubenswrapper[4742]: I1004 03:25:08.320125 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g8xs6" Oct 04 03:25:08 crc kubenswrapper[4742]: I1004 03:25:08.349235 4742 scope.go:117] "RemoveContainer" containerID="2ff9c14ac06b395d1afa2473e2ad4cb6b08d5e24b3e13f1f56dee259e2fd7004" Oct 04 03:25:08 crc kubenswrapper[4742]: I1004 03:25:08.391517 4742 scope.go:117] "RemoveContainer" containerID="702850a871b6e344b5926af819a8b44aab11c822d73185f526ab86968ff3039e" Oct 04 03:25:08 crc kubenswrapper[4742]: I1004 03:25:08.414701 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g8xs6"] Oct 04 03:25:08 crc kubenswrapper[4742]: I1004 03:25:08.419764 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-g8xs6"] Oct 04 03:25:08 crc kubenswrapper[4742]: I1004 03:25:08.866806 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c44b7e8b-6f70-49ff-9be1-857ff2209913" path="/var/lib/kubelet/pods/c44b7e8b-6f70-49ff-9be1-857ff2209913/volumes" Oct 04 03:25:13 crc kubenswrapper[4742]: I1004 03:25:13.209971 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wgl5f"] Oct 04 03:25:13 crc kubenswrapper[4742]: E1004 03:25:13.214619 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c44b7e8b-6f70-49ff-9be1-857ff2209913" containerName="extract-content" Oct 04 03:25:13 crc kubenswrapper[4742]: I1004 03:25:13.214665 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="c44b7e8b-6f70-49ff-9be1-857ff2209913" containerName="extract-content" Oct 04 03:25:13 crc kubenswrapper[4742]: E1004 03:25:13.214689 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c44b7e8b-6f70-49ff-9be1-857ff2209913" containerName="registry-server" Oct 04 03:25:13 crc kubenswrapper[4742]: I1004 03:25:13.214700 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="c44b7e8b-6f70-49ff-9be1-857ff2209913" containerName="registry-server" Oct 04 03:25:13 crc kubenswrapper[4742]: E1004 03:25:13.214716 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c44b7e8b-6f70-49ff-9be1-857ff2209913" containerName="extract-utilities" Oct 04 03:25:13 crc kubenswrapper[4742]: I1004 03:25:13.214729 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="c44b7e8b-6f70-49ff-9be1-857ff2209913" containerName="extract-utilities" Oct 04 03:25:13 crc kubenswrapper[4742]: I1004 03:25:13.214909 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="c44b7e8b-6f70-49ff-9be1-857ff2209913" containerName="registry-server" Oct 04 03:25:13 crc kubenswrapper[4742]: I1004 03:25:13.216220 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wgl5f" Oct 04 03:25:13 crc kubenswrapper[4742]: I1004 03:25:13.226947 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wgl5f"] Oct 04 03:25:13 crc kubenswrapper[4742]: I1004 03:25:13.316248 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71215cf9-91af-432e-ac53-926c80568fc0-utilities\") pod \"redhat-marketplace-wgl5f\" (UID: \"71215cf9-91af-432e-ac53-926c80568fc0\") " pod="openshift-marketplace/redhat-marketplace-wgl5f" Oct 04 03:25:13 crc kubenswrapper[4742]: I1004 03:25:13.316356 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71215cf9-91af-432e-ac53-926c80568fc0-catalog-content\") pod \"redhat-marketplace-wgl5f\" (UID: \"71215cf9-91af-432e-ac53-926c80568fc0\") " pod="openshift-marketplace/redhat-marketplace-wgl5f" Oct 04 03:25:13 crc kubenswrapper[4742]: I1004 03:25:13.316428 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9zw8\" (UniqueName: \"kubernetes.io/projected/71215cf9-91af-432e-ac53-926c80568fc0-kube-api-access-z9zw8\") pod \"redhat-marketplace-wgl5f\" (UID: \"71215cf9-91af-432e-ac53-926c80568fc0\") " pod="openshift-marketplace/redhat-marketplace-wgl5f" Oct 04 03:25:13 crc kubenswrapper[4742]: I1004 03:25:13.417734 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71215cf9-91af-432e-ac53-926c80568fc0-catalog-content\") pod \"redhat-marketplace-wgl5f\" (UID: \"71215cf9-91af-432e-ac53-926c80568fc0\") " pod="openshift-marketplace/redhat-marketplace-wgl5f" Oct 04 03:25:13 crc kubenswrapper[4742]: I1004 03:25:13.417880 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9zw8\" (UniqueName: \"kubernetes.io/projected/71215cf9-91af-432e-ac53-926c80568fc0-kube-api-access-z9zw8\") pod \"redhat-marketplace-wgl5f\" (UID: \"71215cf9-91af-432e-ac53-926c80568fc0\") " pod="openshift-marketplace/redhat-marketplace-wgl5f" Oct 04 03:25:13 crc kubenswrapper[4742]: I1004 03:25:13.417979 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71215cf9-91af-432e-ac53-926c80568fc0-utilities\") pod \"redhat-marketplace-wgl5f\" (UID: \"71215cf9-91af-432e-ac53-926c80568fc0\") " pod="openshift-marketplace/redhat-marketplace-wgl5f" Oct 04 03:25:13 crc kubenswrapper[4742]: I1004 03:25:13.418962 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71215cf9-91af-432e-ac53-926c80568fc0-utilities\") pod \"redhat-marketplace-wgl5f\" (UID: \"71215cf9-91af-432e-ac53-926c80568fc0\") " pod="openshift-marketplace/redhat-marketplace-wgl5f" Oct 04 03:25:13 crc kubenswrapper[4742]: I1004 03:25:13.419007 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71215cf9-91af-432e-ac53-926c80568fc0-catalog-content\") pod \"redhat-marketplace-wgl5f\" (UID: \"71215cf9-91af-432e-ac53-926c80568fc0\") " pod="openshift-marketplace/redhat-marketplace-wgl5f" Oct 04 03:25:13 crc kubenswrapper[4742]: I1004 03:25:13.447206 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9zw8\" (UniqueName: \"kubernetes.io/projected/71215cf9-91af-432e-ac53-926c80568fc0-kube-api-access-z9zw8\") pod \"redhat-marketplace-wgl5f\" (UID: \"71215cf9-91af-432e-ac53-926c80568fc0\") " pod="openshift-marketplace/redhat-marketplace-wgl5f" Oct 04 03:25:13 crc kubenswrapper[4742]: I1004 03:25:13.548596 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wgl5f" Oct 04 03:25:13 crc kubenswrapper[4742]: I1004 03:25:13.902605 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wgl5f"] Oct 04 03:25:14 crc kubenswrapper[4742]: I1004 03:25:14.373878 4742 generic.go:334] "Generic (PLEG): container finished" podID="71215cf9-91af-432e-ac53-926c80568fc0" containerID="11a0bd0117c082b3a097aa6cdb5b8c1d84700823ebb45bd8fd5b69536b6545b3" exitCode=0 Oct 04 03:25:14 crc kubenswrapper[4742]: I1004 03:25:14.373925 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wgl5f" event={"ID":"71215cf9-91af-432e-ac53-926c80568fc0","Type":"ContainerDied","Data":"11a0bd0117c082b3a097aa6cdb5b8c1d84700823ebb45bd8fd5b69536b6545b3"} Oct 04 03:25:14 crc kubenswrapper[4742]: I1004 03:25:14.373951 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wgl5f" event={"ID":"71215cf9-91af-432e-ac53-926c80568fc0","Type":"ContainerStarted","Data":"d35e09374e206bf2230895229eb629a52de3082947cb3ff3c0b20e95db59643a"} Oct 04 03:25:14 crc kubenswrapper[4742]: I1004 03:25:14.398765 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-index-2sn2g"] Oct 04 03:25:14 crc kubenswrapper[4742]: I1004 03:25:14.400205 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-2sn2g" Oct 04 03:25:14 crc kubenswrapper[4742]: I1004 03:25:14.404548 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-index-dockercfg-gl6lf" Oct 04 03:25:14 crc kubenswrapper[4742]: I1004 03:25:14.418048 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-2sn2g"] Oct 04 03:25:14 crc kubenswrapper[4742]: I1004 03:25:14.540743 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qds64\" (UniqueName: \"kubernetes.io/projected/d574cbce-5f5f-4093-8e44-4bd06c8c9eae-kube-api-access-qds64\") pod \"horizon-operator-index-2sn2g\" (UID: \"d574cbce-5f5f-4093-8e44-4bd06c8c9eae\") " pod="openstack-operators/horizon-operator-index-2sn2g" Oct 04 03:25:14 crc kubenswrapper[4742]: I1004 03:25:14.642077 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qds64\" (UniqueName: \"kubernetes.io/projected/d574cbce-5f5f-4093-8e44-4bd06c8c9eae-kube-api-access-qds64\") pod \"horizon-operator-index-2sn2g\" (UID: \"d574cbce-5f5f-4093-8e44-4bd06c8c9eae\") " pod="openstack-operators/horizon-operator-index-2sn2g" Oct 04 03:25:14 crc kubenswrapper[4742]: I1004 03:25:14.664223 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qds64\" (UniqueName: \"kubernetes.io/projected/d574cbce-5f5f-4093-8e44-4bd06c8c9eae-kube-api-access-qds64\") pod \"horizon-operator-index-2sn2g\" (UID: \"d574cbce-5f5f-4093-8e44-4bd06c8c9eae\") " pod="openstack-operators/horizon-operator-index-2sn2g" Oct 04 03:25:14 crc kubenswrapper[4742]: I1004 03:25:14.720043 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-2sn2g" Oct 04 03:25:15 crc kubenswrapper[4742]: I1004 03:25:15.076825 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-2sn2g"] Oct 04 03:25:15 crc kubenswrapper[4742]: I1004 03:25:15.388843 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-2sn2g" event={"ID":"d574cbce-5f5f-4093-8e44-4bd06c8c9eae","Type":"ContainerStarted","Data":"184ccf46c169ba2ecf6d43f4e8e6c52679203a23e1100283543a0b3a5e45366c"} Oct 04 03:25:16 crc kubenswrapper[4742]: I1004 03:25:16.408153 4742 generic.go:334] "Generic (PLEG): container finished" podID="71215cf9-91af-432e-ac53-926c80568fc0" containerID="bdbcd02b5ae4fb312a43e242c7f398d4d9e48705eb367016f214e641fa0504ee" exitCode=0 Oct 04 03:25:16 crc kubenswrapper[4742]: I1004 03:25:16.408292 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wgl5f" event={"ID":"71215cf9-91af-432e-ac53-926c80568fc0","Type":"ContainerDied","Data":"bdbcd02b5ae4fb312a43e242c7f398d4d9e48705eb367016f214e641fa0504ee"} Oct 04 03:25:17 crc kubenswrapper[4742]: I1004 03:25:17.418954 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wgl5f" event={"ID":"71215cf9-91af-432e-ac53-926c80568fc0","Type":"ContainerStarted","Data":"e0959254824de6d67a03c6cca46fdc8d7861539e92ee5569fe817dbce20c4a1a"} Oct 04 03:25:17 crc kubenswrapper[4742]: I1004 03:25:17.420732 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-2sn2g" event={"ID":"d574cbce-5f5f-4093-8e44-4bd06c8c9eae","Type":"ContainerStarted","Data":"300448a688a6b2c164146739f4d5ba4f9cb75dbd3c789fefa65b88bbae7171a0"} Oct 04 03:25:17 crc kubenswrapper[4742]: I1004 03:25:17.438621 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wgl5f" podStartSLOduration=1.740278133 podStartE2EDuration="4.438598239s" podCreationTimestamp="2025-10-04 03:25:13 +0000 UTC" firstStartedPulling="2025-10-04 03:25:14.375453567 +0000 UTC m=+880.307790743" lastFinishedPulling="2025-10-04 03:25:17.073773673 +0000 UTC m=+883.006110849" observedRunningTime="2025-10-04 03:25:17.437441289 +0000 UTC m=+883.369778465" watchObservedRunningTime="2025-10-04 03:25:17.438598239 +0000 UTC m=+883.370935435" Oct 04 03:25:17 crc kubenswrapper[4742]: I1004 03:25:17.451915 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-index-2sn2g" podStartSLOduration=1.48134116 podStartE2EDuration="3.451895272s" podCreationTimestamp="2025-10-04 03:25:14 +0000 UTC" firstStartedPulling="2025-10-04 03:25:15.104648348 +0000 UTC m=+881.036985554" lastFinishedPulling="2025-10-04 03:25:17.07520248 +0000 UTC m=+883.007539666" observedRunningTime="2025-10-04 03:25:17.451196144 +0000 UTC m=+883.383533330" watchObservedRunningTime="2025-10-04 03:25:17.451895272 +0000 UTC m=+883.384232458" Oct 04 03:25:19 crc kubenswrapper[4742]: I1004 03:25:19.606101 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-index-gppbc"] Oct 04 03:25:19 crc kubenswrapper[4742]: I1004 03:25:19.607687 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-gppbc" Oct 04 03:25:19 crc kubenswrapper[4742]: I1004 03:25:19.610225 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-index-dockercfg-zckjm" Oct 04 03:25:19 crc kubenswrapper[4742]: I1004 03:25:19.615739 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-gppbc"] Oct 04 03:25:19 crc kubenswrapper[4742]: I1004 03:25:19.732646 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pc7dv\" (UniqueName: \"kubernetes.io/projected/a7fdeeb5-30fc-42da-b8c6-f5e91634cf91-kube-api-access-pc7dv\") pod \"swift-operator-index-gppbc\" (UID: \"a7fdeeb5-30fc-42da-b8c6-f5e91634cf91\") " pod="openstack-operators/swift-operator-index-gppbc" Oct 04 03:25:19 crc kubenswrapper[4742]: I1004 03:25:19.834777 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pc7dv\" (UniqueName: \"kubernetes.io/projected/a7fdeeb5-30fc-42da-b8c6-f5e91634cf91-kube-api-access-pc7dv\") pod \"swift-operator-index-gppbc\" (UID: \"a7fdeeb5-30fc-42da-b8c6-f5e91634cf91\") " pod="openstack-operators/swift-operator-index-gppbc" Oct 04 03:25:19 crc kubenswrapper[4742]: I1004 03:25:19.860482 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pc7dv\" (UniqueName: \"kubernetes.io/projected/a7fdeeb5-30fc-42da-b8c6-f5e91634cf91-kube-api-access-pc7dv\") pod \"swift-operator-index-gppbc\" (UID: \"a7fdeeb5-30fc-42da-b8c6-f5e91634cf91\") " pod="openstack-operators/swift-operator-index-gppbc" Oct 04 03:25:19 crc kubenswrapper[4742]: I1004 03:25:19.969162 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-gppbc" Oct 04 03:25:20 crc kubenswrapper[4742]: I1004 03:25:20.274401 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-gppbc"] Oct 04 03:25:20 crc kubenswrapper[4742]: I1004 03:25:20.448301 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-gppbc" event={"ID":"a7fdeeb5-30fc-42da-b8c6-f5e91634cf91","Type":"ContainerStarted","Data":"db43eee66fed9b09a7563986edbf1c85395c810c86924980a01ab9fb64a24355"} Oct 04 03:25:20 crc kubenswrapper[4742]: I1004 03:25:20.876109 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-db-create-vt8sv"] Oct 04 03:25:20 crc kubenswrapper[4742]: I1004 03:25:20.877850 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-vt8sv" Oct 04 03:25:20 crc kubenswrapper[4742]: I1004 03:25:20.881865 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-create-vt8sv"] Oct 04 03:25:20 crc kubenswrapper[4742]: I1004 03:25:20.955426 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pr4kn\" (UniqueName: \"kubernetes.io/projected/9f409680-b99e-4ffb-bc3c-7f69c8856347-kube-api-access-pr4kn\") pod \"keystone-db-create-vt8sv\" (UID: \"9f409680-b99e-4ffb-bc3c-7f69c8856347\") " pod="glance-kuttl-tests/keystone-db-create-vt8sv" Oct 04 03:25:21 crc kubenswrapper[4742]: I1004 03:25:21.057097 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pr4kn\" (UniqueName: \"kubernetes.io/projected/9f409680-b99e-4ffb-bc3c-7f69c8856347-kube-api-access-pr4kn\") pod \"keystone-db-create-vt8sv\" (UID: \"9f409680-b99e-4ffb-bc3c-7f69c8856347\") " pod="glance-kuttl-tests/keystone-db-create-vt8sv" Oct 04 03:25:21 crc kubenswrapper[4742]: I1004 03:25:21.094822 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pr4kn\" (UniqueName: \"kubernetes.io/projected/9f409680-b99e-4ffb-bc3c-7f69c8856347-kube-api-access-pr4kn\") pod \"keystone-db-create-vt8sv\" (UID: \"9f409680-b99e-4ffb-bc3c-7f69c8856347\") " pod="glance-kuttl-tests/keystone-db-create-vt8sv" Oct 04 03:25:21 crc kubenswrapper[4742]: I1004 03:25:21.213715 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-vt8sv" Oct 04 03:25:21 crc kubenswrapper[4742]: I1004 03:25:21.515628 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 03:25:21 crc kubenswrapper[4742]: I1004 03:25:21.672702 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-create-vt8sv"] Oct 04 03:25:22 crc kubenswrapper[4742]: I1004 03:25:22.465663 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-gppbc" event={"ID":"a7fdeeb5-30fc-42da-b8c6-f5e91634cf91","Type":"ContainerStarted","Data":"1ca5f4969958115331f464e829e502f5a34d9760864331d81210009f6eb5a27f"} Oct 04 03:25:22 crc kubenswrapper[4742]: I1004 03:25:22.470028 4742 generic.go:334] "Generic (PLEG): container finished" podID="9f409680-b99e-4ffb-bc3c-7f69c8856347" containerID="ba39dafe40c73aa4f142a272463e0fc3da2dd8339ff54fd83a70b4cffc511d95" exitCode=0 Oct 04 03:25:22 crc kubenswrapper[4742]: I1004 03:25:22.470099 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-vt8sv" event={"ID":"9f409680-b99e-4ffb-bc3c-7f69c8856347","Type":"ContainerDied","Data":"ba39dafe40c73aa4f142a272463e0fc3da2dd8339ff54fd83a70b4cffc511d95"} Oct 04 03:25:22 crc kubenswrapper[4742]: I1004 03:25:22.470179 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-vt8sv" event={"ID":"9f409680-b99e-4ffb-bc3c-7f69c8856347","Type":"ContainerStarted","Data":"81abd631e11fceec3d9bd5c54c67d5da837d9964cbbe339b187fd809dc4ae5ad"} Oct 04 03:25:22 crc kubenswrapper[4742]: I1004 03:25:22.485214 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-index-gppbc" podStartSLOduration=1.855144503 podStartE2EDuration="3.485190116s" podCreationTimestamp="2025-10-04 03:25:19 +0000 UTC" firstStartedPulling="2025-10-04 03:25:20.281978941 +0000 UTC m=+886.214316127" lastFinishedPulling="2025-10-04 03:25:21.912024564 +0000 UTC m=+887.844361740" observedRunningTime="2025-10-04 03:25:22.481608184 +0000 UTC m=+888.413945370" watchObservedRunningTime="2025-10-04 03:25:22.485190116 +0000 UTC m=+888.417527322" Oct 04 03:25:23 crc kubenswrapper[4742]: I1004 03:25:23.549706 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wgl5f" Oct 04 03:25:23 crc kubenswrapper[4742]: I1004 03:25:23.550053 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wgl5f" Oct 04 03:25:23 crc kubenswrapper[4742]: I1004 03:25:23.614599 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wgl5f" Oct 04 03:25:23 crc kubenswrapper[4742]: I1004 03:25:23.831236 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-vt8sv" Oct 04 03:25:23 crc kubenswrapper[4742]: I1004 03:25:23.903186 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pr4kn\" (UniqueName: \"kubernetes.io/projected/9f409680-b99e-4ffb-bc3c-7f69c8856347-kube-api-access-pr4kn\") pod \"9f409680-b99e-4ffb-bc3c-7f69c8856347\" (UID: \"9f409680-b99e-4ffb-bc3c-7f69c8856347\") " Oct 04 03:25:23 crc kubenswrapper[4742]: I1004 03:25:23.912168 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f409680-b99e-4ffb-bc3c-7f69c8856347-kube-api-access-pr4kn" (OuterVolumeSpecName: "kube-api-access-pr4kn") pod "9f409680-b99e-4ffb-bc3c-7f69c8856347" (UID: "9f409680-b99e-4ffb-bc3c-7f69c8856347"). InnerVolumeSpecName "kube-api-access-pr4kn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:24 crc kubenswrapper[4742]: I1004 03:25:24.006056 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pr4kn\" (UniqueName: \"kubernetes.io/projected/9f409680-b99e-4ffb-bc3c-7f69c8856347-kube-api-access-pr4kn\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:24 crc kubenswrapper[4742]: I1004 03:25:24.490641 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-vt8sv" Oct 04 03:25:24 crc kubenswrapper[4742]: I1004 03:25:24.491729 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-vt8sv" event={"ID":"9f409680-b99e-4ffb-bc3c-7f69c8856347","Type":"ContainerDied","Data":"81abd631e11fceec3d9bd5c54c67d5da837d9964cbbe339b187fd809dc4ae5ad"} Oct 04 03:25:24 crc kubenswrapper[4742]: I1004 03:25:24.491802 4742 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="81abd631e11fceec3d9bd5c54c67d5da837d9964cbbe339b187fd809dc4ae5ad" Oct 04 03:25:24 crc kubenswrapper[4742]: I1004 03:25:24.580432 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wgl5f" Oct 04 03:25:24 crc kubenswrapper[4742]: I1004 03:25:24.720969 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-index-2sn2g" Oct 04 03:25:24 crc kubenswrapper[4742]: I1004 03:25:24.721030 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/horizon-operator-index-2sn2g" Oct 04 03:25:24 crc kubenswrapper[4742]: I1004 03:25:24.766700 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/horizon-operator-index-2sn2g" Oct 04 03:25:25 crc kubenswrapper[4742]: I1004 03:25:25.547410 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-index-2sn2g" Oct 04 03:25:27 crc kubenswrapper[4742]: I1004 03:25:27.794489 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wgl5f"] Oct 04 03:25:27 crc kubenswrapper[4742]: I1004 03:25:27.796390 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wgl5f" podUID="71215cf9-91af-432e-ac53-926c80568fc0" containerName="registry-server" containerID="cri-o://e0959254824de6d67a03c6cca46fdc8d7861539e92ee5569fe817dbce20c4a1a" gracePeriod=2 Oct 04 03:25:28 crc kubenswrapper[4742]: I1004 03:25:28.213201 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wgl5f" Oct 04 03:25:28 crc kubenswrapper[4742]: I1004 03:25:28.276079 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9zw8\" (UniqueName: \"kubernetes.io/projected/71215cf9-91af-432e-ac53-926c80568fc0-kube-api-access-z9zw8\") pod \"71215cf9-91af-432e-ac53-926c80568fc0\" (UID: \"71215cf9-91af-432e-ac53-926c80568fc0\") " Oct 04 03:25:28 crc kubenswrapper[4742]: I1004 03:25:28.276219 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71215cf9-91af-432e-ac53-926c80568fc0-catalog-content\") pod \"71215cf9-91af-432e-ac53-926c80568fc0\" (UID: \"71215cf9-91af-432e-ac53-926c80568fc0\") " Oct 04 03:25:28 crc kubenswrapper[4742]: I1004 03:25:28.276686 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71215cf9-91af-432e-ac53-926c80568fc0-utilities\") pod \"71215cf9-91af-432e-ac53-926c80568fc0\" (UID: \"71215cf9-91af-432e-ac53-926c80568fc0\") " Oct 04 03:25:28 crc kubenswrapper[4742]: I1004 03:25:28.278497 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71215cf9-91af-432e-ac53-926c80568fc0-utilities" (OuterVolumeSpecName: "utilities") pod "71215cf9-91af-432e-ac53-926c80568fc0" (UID: "71215cf9-91af-432e-ac53-926c80568fc0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:25:28 crc kubenswrapper[4742]: I1004 03:25:28.285433 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71215cf9-91af-432e-ac53-926c80568fc0-kube-api-access-z9zw8" (OuterVolumeSpecName: "kube-api-access-z9zw8") pod "71215cf9-91af-432e-ac53-926c80568fc0" (UID: "71215cf9-91af-432e-ac53-926c80568fc0"). InnerVolumeSpecName "kube-api-access-z9zw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:28 crc kubenswrapper[4742]: I1004 03:25:28.295468 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71215cf9-91af-432e-ac53-926c80568fc0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "71215cf9-91af-432e-ac53-926c80568fc0" (UID: "71215cf9-91af-432e-ac53-926c80568fc0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:25:28 crc kubenswrapper[4742]: I1004 03:25:28.379646 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9zw8\" (UniqueName: \"kubernetes.io/projected/71215cf9-91af-432e-ac53-926c80568fc0-kube-api-access-z9zw8\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:28 crc kubenswrapper[4742]: I1004 03:25:28.379692 4742 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71215cf9-91af-432e-ac53-926c80568fc0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:28 crc kubenswrapper[4742]: I1004 03:25:28.379704 4742 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71215cf9-91af-432e-ac53-926c80568fc0-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:28 crc kubenswrapper[4742]: I1004 03:25:28.533626 4742 generic.go:334] "Generic (PLEG): container finished" podID="71215cf9-91af-432e-ac53-926c80568fc0" containerID="e0959254824de6d67a03c6cca46fdc8d7861539e92ee5569fe817dbce20c4a1a" exitCode=0 Oct 04 03:25:28 crc kubenswrapper[4742]: I1004 03:25:28.533844 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wgl5f" event={"ID":"71215cf9-91af-432e-ac53-926c80568fc0","Type":"ContainerDied","Data":"e0959254824de6d67a03c6cca46fdc8d7861539e92ee5569fe817dbce20c4a1a"} Oct 04 03:25:28 crc kubenswrapper[4742]: I1004 03:25:28.534178 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wgl5f" event={"ID":"71215cf9-91af-432e-ac53-926c80568fc0","Type":"ContainerDied","Data":"d35e09374e206bf2230895229eb629a52de3082947cb3ff3c0b20e95db59643a"} Oct 04 03:25:28 crc kubenswrapper[4742]: I1004 03:25:28.533966 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wgl5f" Oct 04 03:25:28 crc kubenswrapper[4742]: I1004 03:25:28.534235 4742 scope.go:117] "RemoveContainer" containerID="e0959254824de6d67a03c6cca46fdc8d7861539e92ee5569fe817dbce20c4a1a" Oct 04 03:25:28 crc kubenswrapper[4742]: I1004 03:25:28.575368 4742 scope.go:117] "RemoveContainer" containerID="bdbcd02b5ae4fb312a43e242c7f398d4d9e48705eb367016f214e641fa0504ee" Oct 04 03:25:28 crc kubenswrapper[4742]: I1004 03:25:28.582976 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wgl5f"] Oct 04 03:25:28 crc kubenswrapper[4742]: I1004 03:25:28.591066 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wgl5f"] Oct 04 03:25:28 crc kubenswrapper[4742]: I1004 03:25:28.601706 4742 scope.go:117] "RemoveContainer" containerID="11a0bd0117c082b3a097aa6cdb5b8c1d84700823ebb45bd8fd5b69536b6545b3" Oct 04 03:25:28 crc kubenswrapper[4742]: I1004 03:25:28.643114 4742 scope.go:117] "RemoveContainer" containerID="e0959254824de6d67a03c6cca46fdc8d7861539e92ee5569fe817dbce20c4a1a" Oct 04 03:25:28 crc kubenswrapper[4742]: E1004 03:25:28.643882 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0959254824de6d67a03c6cca46fdc8d7861539e92ee5569fe817dbce20c4a1a\": container with ID starting with e0959254824de6d67a03c6cca46fdc8d7861539e92ee5569fe817dbce20c4a1a not found: ID does not exist" containerID="e0959254824de6d67a03c6cca46fdc8d7861539e92ee5569fe817dbce20c4a1a" Oct 04 03:25:28 crc kubenswrapper[4742]: I1004 03:25:28.643952 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0959254824de6d67a03c6cca46fdc8d7861539e92ee5569fe817dbce20c4a1a"} err="failed to get container status \"e0959254824de6d67a03c6cca46fdc8d7861539e92ee5569fe817dbce20c4a1a\": rpc error: code = NotFound desc = could not find container \"e0959254824de6d67a03c6cca46fdc8d7861539e92ee5569fe817dbce20c4a1a\": container with ID starting with e0959254824de6d67a03c6cca46fdc8d7861539e92ee5569fe817dbce20c4a1a not found: ID does not exist" Oct 04 03:25:28 crc kubenswrapper[4742]: I1004 03:25:28.643995 4742 scope.go:117] "RemoveContainer" containerID="bdbcd02b5ae4fb312a43e242c7f398d4d9e48705eb367016f214e641fa0504ee" Oct 04 03:25:28 crc kubenswrapper[4742]: E1004 03:25:28.644641 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdbcd02b5ae4fb312a43e242c7f398d4d9e48705eb367016f214e641fa0504ee\": container with ID starting with bdbcd02b5ae4fb312a43e242c7f398d4d9e48705eb367016f214e641fa0504ee not found: ID does not exist" containerID="bdbcd02b5ae4fb312a43e242c7f398d4d9e48705eb367016f214e641fa0504ee" Oct 04 03:25:28 crc kubenswrapper[4742]: I1004 03:25:28.644723 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdbcd02b5ae4fb312a43e242c7f398d4d9e48705eb367016f214e641fa0504ee"} err="failed to get container status \"bdbcd02b5ae4fb312a43e242c7f398d4d9e48705eb367016f214e641fa0504ee\": rpc error: code = NotFound desc = could not find container \"bdbcd02b5ae4fb312a43e242c7f398d4d9e48705eb367016f214e641fa0504ee\": container with ID starting with bdbcd02b5ae4fb312a43e242c7f398d4d9e48705eb367016f214e641fa0504ee not found: ID does not exist" Oct 04 03:25:28 crc kubenswrapper[4742]: I1004 03:25:28.644773 4742 scope.go:117] "RemoveContainer" containerID="11a0bd0117c082b3a097aa6cdb5b8c1d84700823ebb45bd8fd5b69536b6545b3" Oct 04 03:25:28 crc kubenswrapper[4742]: E1004 03:25:28.645214 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11a0bd0117c082b3a097aa6cdb5b8c1d84700823ebb45bd8fd5b69536b6545b3\": container with ID starting with 11a0bd0117c082b3a097aa6cdb5b8c1d84700823ebb45bd8fd5b69536b6545b3 not found: ID does not exist" containerID="11a0bd0117c082b3a097aa6cdb5b8c1d84700823ebb45bd8fd5b69536b6545b3" Oct 04 03:25:28 crc kubenswrapper[4742]: I1004 03:25:28.645303 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11a0bd0117c082b3a097aa6cdb5b8c1d84700823ebb45bd8fd5b69536b6545b3"} err="failed to get container status \"11a0bd0117c082b3a097aa6cdb5b8c1d84700823ebb45bd8fd5b69536b6545b3\": rpc error: code = NotFound desc = could not find container \"11a0bd0117c082b3a097aa6cdb5b8c1d84700823ebb45bd8fd5b69536b6545b3\": container with ID starting with 11a0bd0117c082b3a097aa6cdb5b8c1d84700823ebb45bd8fd5b69536b6545b3 not found: ID does not exist" Oct 04 03:25:28 crc kubenswrapper[4742]: I1004 03:25:28.866224 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71215cf9-91af-432e-ac53-926c80568fc0" path="/var/lib/kubelet/pods/71215cf9-91af-432e-ac53-926c80568fc0/volumes" Oct 04 03:25:29 crc kubenswrapper[4742]: I1004 03:25:29.970495 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/swift-operator-index-gppbc" Oct 04 03:25:29 crc kubenswrapper[4742]: I1004 03:25:29.971149 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-index-gppbc" Oct 04 03:25:30 crc kubenswrapper[4742]: I1004 03:25:30.023510 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/swift-operator-index-gppbc" Oct 04 03:25:30 crc kubenswrapper[4742]: I1004 03:25:30.614836 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-index-gppbc" Oct 04 03:25:30 crc kubenswrapper[4742]: I1004 03:25:30.758590 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-f46e-account-create-88lml"] Oct 04 03:25:30 crc kubenswrapper[4742]: E1004 03:25:30.759037 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f409680-b99e-4ffb-bc3c-7f69c8856347" containerName="mariadb-database-create" Oct 04 03:25:30 crc kubenswrapper[4742]: I1004 03:25:30.759056 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f409680-b99e-4ffb-bc3c-7f69c8856347" containerName="mariadb-database-create" Oct 04 03:25:30 crc kubenswrapper[4742]: E1004 03:25:30.759071 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71215cf9-91af-432e-ac53-926c80568fc0" containerName="extract-content" Oct 04 03:25:30 crc kubenswrapper[4742]: I1004 03:25:30.759080 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="71215cf9-91af-432e-ac53-926c80568fc0" containerName="extract-content" Oct 04 03:25:30 crc kubenswrapper[4742]: E1004 03:25:30.759090 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71215cf9-91af-432e-ac53-926c80568fc0" containerName="registry-server" Oct 04 03:25:30 crc kubenswrapper[4742]: I1004 03:25:30.759098 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="71215cf9-91af-432e-ac53-926c80568fc0" containerName="registry-server" Oct 04 03:25:30 crc kubenswrapper[4742]: E1004 03:25:30.759114 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71215cf9-91af-432e-ac53-926c80568fc0" containerName="extract-utilities" Oct 04 03:25:30 crc kubenswrapper[4742]: I1004 03:25:30.759121 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="71215cf9-91af-432e-ac53-926c80568fc0" containerName="extract-utilities" Oct 04 03:25:30 crc kubenswrapper[4742]: I1004 03:25:30.759259 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f409680-b99e-4ffb-bc3c-7f69c8856347" containerName="mariadb-database-create" Oct 04 03:25:30 crc kubenswrapper[4742]: I1004 03:25:30.759290 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="71215cf9-91af-432e-ac53-926c80568fc0" containerName="registry-server" Oct 04 03:25:30 crc kubenswrapper[4742]: I1004 03:25:30.759934 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-f46e-account-create-88lml" Oct 04 03:25:30 crc kubenswrapper[4742]: I1004 03:25:30.763577 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-db-secret" Oct 04 03:25:30 crc kubenswrapper[4742]: I1004 03:25:30.779512 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-f46e-account-create-88lml"] Oct 04 03:25:30 crc kubenswrapper[4742]: I1004 03:25:30.815679 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmsqq\" (UniqueName: \"kubernetes.io/projected/0af4aaa9-9a74-427f-9940-c5ebcf1d3efb-kube-api-access-nmsqq\") pod \"keystone-f46e-account-create-88lml\" (UID: \"0af4aaa9-9a74-427f-9940-c5ebcf1d3efb\") " pod="glance-kuttl-tests/keystone-f46e-account-create-88lml" Oct 04 03:25:30 crc kubenswrapper[4742]: I1004 03:25:30.917443 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmsqq\" (UniqueName: \"kubernetes.io/projected/0af4aaa9-9a74-427f-9940-c5ebcf1d3efb-kube-api-access-nmsqq\") pod \"keystone-f46e-account-create-88lml\" (UID: \"0af4aaa9-9a74-427f-9940-c5ebcf1d3efb\") " pod="glance-kuttl-tests/keystone-f46e-account-create-88lml" Oct 04 03:25:30 crc kubenswrapper[4742]: I1004 03:25:30.945418 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmsqq\" (UniqueName: \"kubernetes.io/projected/0af4aaa9-9a74-427f-9940-c5ebcf1d3efb-kube-api-access-nmsqq\") pod \"keystone-f46e-account-create-88lml\" (UID: \"0af4aaa9-9a74-427f-9940-c5ebcf1d3efb\") " pod="glance-kuttl-tests/keystone-f46e-account-create-88lml" Oct 04 03:25:31 crc kubenswrapper[4742]: I1004 03:25:31.097443 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-f46e-account-create-88lml" Oct 04 03:25:31 crc kubenswrapper[4742]: I1004 03:25:31.392758 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-f46e-account-create-88lml"] Oct 04 03:25:31 crc kubenswrapper[4742]: I1004 03:25:31.564636 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-f46e-account-create-88lml" event={"ID":"0af4aaa9-9a74-427f-9940-c5ebcf1d3efb","Type":"ContainerStarted","Data":"13d43bf598a44a08017a561c7497d23b029d074daf71f7a45a1cb4beb307eb22"} Oct 04 03:25:32 crc kubenswrapper[4742]: I1004 03:25:32.578083 4742 generic.go:334] "Generic (PLEG): container finished" podID="0af4aaa9-9a74-427f-9940-c5ebcf1d3efb" containerID="fc45ed162295579347770159149bc120feb68223e97a9ae51a67c5fb8ec22635" exitCode=0 Oct 04 03:25:32 crc kubenswrapper[4742]: I1004 03:25:32.578200 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-f46e-account-create-88lml" event={"ID":"0af4aaa9-9a74-427f-9940-c5ebcf1d3efb","Type":"ContainerDied","Data":"fc45ed162295579347770159149bc120feb68223e97a9ae51a67c5fb8ec22635"} Oct 04 03:25:33 crc kubenswrapper[4742]: I1004 03:25:33.888737 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-f46e-account-create-88lml" Oct 04 03:25:33 crc kubenswrapper[4742]: I1004 03:25:33.961791 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmsqq\" (UniqueName: \"kubernetes.io/projected/0af4aaa9-9a74-427f-9940-c5ebcf1d3efb-kube-api-access-nmsqq\") pod \"0af4aaa9-9a74-427f-9940-c5ebcf1d3efb\" (UID: \"0af4aaa9-9a74-427f-9940-c5ebcf1d3efb\") " Oct 04 03:25:33 crc kubenswrapper[4742]: I1004 03:25:33.969404 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0af4aaa9-9a74-427f-9940-c5ebcf1d3efb-kube-api-access-nmsqq" (OuterVolumeSpecName: "kube-api-access-nmsqq") pod "0af4aaa9-9a74-427f-9940-c5ebcf1d3efb" (UID: "0af4aaa9-9a74-427f-9940-c5ebcf1d3efb"). InnerVolumeSpecName "kube-api-access-nmsqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:34 crc kubenswrapper[4742]: I1004 03:25:34.063598 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmsqq\" (UniqueName: \"kubernetes.io/projected/0af4aaa9-9a74-427f-9940-c5ebcf1d3efb-kube-api-access-nmsqq\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:34 crc kubenswrapper[4742]: I1004 03:25:34.598964 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-f46e-account-create-88lml" event={"ID":"0af4aaa9-9a74-427f-9940-c5ebcf1d3efb","Type":"ContainerDied","Data":"13d43bf598a44a08017a561c7497d23b029d074daf71f7a45a1cb4beb307eb22"} Oct 04 03:25:34 crc kubenswrapper[4742]: I1004 03:25:34.599509 4742 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13d43bf598a44a08017a561c7497d23b029d074daf71f7a45a1cb4beb307eb22" Oct 04 03:25:34 crc kubenswrapper[4742]: I1004 03:25:34.599052 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-f46e-account-create-88lml" Oct 04 03:25:35 crc kubenswrapper[4742]: I1004 03:25:35.654625 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl"] Oct 04 03:25:35 crc kubenswrapper[4742]: E1004 03:25:35.655063 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0af4aaa9-9a74-427f-9940-c5ebcf1d3efb" containerName="mariadb-account-create" Oct 04 03:25:35 crc kubenswrapper[4742]: I1004 03:25:35.655085 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="0af4aaa9-9a74-427f-9940-c5ebcf1d3efb" containerName="mariadb-account-create" Oct 04 03:25:35 crc kubenswrapper[4742]: I1004 03:25:35.655383 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="0af4aaa9-9a74-427f-9940-c5ebcf1d3efb" containerName="mariadb-account-create" Oct 04 03:25:35 crc kubenswrapper[4742]: I1004 03:25:35.657035 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl" Oct 04 03:25:35 crc kubenswrapper[4742]: I1004 03:25:35.659803 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-625gw" Oct 04 03:25:35 crc kubenswrapper[4742]: I1004 03:25:35.670292 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl"] Oct 04 03:25:35 crc kubenswrapper[4742]: I1004 03:25:35.685663 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fda100d8-ce31-46d2-bb94-2edb9aabe58e-bundle\") pod \"c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl\" (UID: \"fda100d8-ce31-46d2-bb94-2edb9aabe58e\") " pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl" Oct 04 03:25:35 crc kubenswrapper[4742]: I1004 03:25:35.685758 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ghm6\" (UniqueName: \"kubernetes.io/projected/fda100d8-ce31-46d2-bb94-2edb9aabe58e-kube-api-access-7ghm6\") pod \"c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl\" (UID: \"fda100d8-ce31-46d2-bb94-2edb9aabe58e\") " pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl" Oct 04 03:25:35 crc kubenswrapper[4742]: I1004 03:25:35.685802 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fda100d8-ce31-46d2-bb94-2edb9aabe58e-util\") pod \"c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl\" (UID: \"fda100d8-ce31-46d2-bb94-2edb9aabe58e\") " pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl" Oct 04 03:25:35 crc kubenswrapper[4742]: I1004 03:25:35.787189 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ghm6\" (UniqueName: \"kubernetes.io/projected/fda100d8-ce31-46d2-bb94-2edb9aabe58e-kube-api-access-7ghm6\") pod \"c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl\" (UID: \"fda100d8-ce31-46d2-bb94-2edb9aabe58e\") " pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl" Oct 04 03:25:35 crc kubenswrapper[4742]: I1004 03:25:35.787281 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fda100d8-ce31-46d2-bb94-2edb9aabe58e-util\") pod \"c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl\" (UID: \"fda100d8-ce31-46d2-bb94-2edb9aabe58e\") " pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl" Oct 04 03:25:35 crc kubenswrapper[4742]: I1004 03:25:35.787352 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fda100d8-ce31-46d2-bb94-2edb9aabe58e-bundle\") pod \"c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl\" (UID: \"fda100d8-ce31-46d2-bb94-2edb9aabe58e\") " pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl" Oct 04 03:25:35 crc kubenswrapper[4742]: I1004 03:25:35.788112 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fda100d8-ce31-46d2-bb94-2edb9aabe58e-bundle\") pod \"c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl\" (UID: \"fda100d8-ce31-46d2-bb94-2edb9aabe58e\") " pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl" Oct 04 03:25:35 crc kubenswrapper[4742]: I1004 03:25:35.788785 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fda100d8-ce31-46d2-bb94-2edb9aabe58e-util\") pod \"c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl\" (UID: \"fda100d8-ce31-46d2-bb94-2edb9aabe58e\") " pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl" Oct 04 03:25:35 crc kubenswrapper[4742]: I1004 03:25:35.825690 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ghm6\" (UniqueName: \"kubernetes.io/projected/fda100d8-ce31-46d2-bb94-2edb9aabe58e-kube-api-access-7ghm6\") pod \"c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl\" (UID: \"fda100d8-ce31-46d2-bb94-2edb9aabe58e\") " pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl" Oct 04 03:25:35 crc kubenswrapper[4742]: I1004 03:25:35.981068 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl" Oct 04 03:25:36 crc kubenswrapper[4742]: I1004 03:25:36.330593 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl"] Oct 04 03:25:36 crc kubenswrapper[4742]: I1004 03:25:36.353057 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-db-sync-j6wks"] Oct 04 03:25:36 crc kubenswrapper[4742]: I1004 03:25:36.354037 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-j6wks" Oct 04 03:25:36 crc kubenswrapper[4742]: I1004 03:25:36.356151 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Oct 04 03:25:36 crc kubenswrapper[4742]: I1004 03:25:36.356290 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-whckd" Oct 04 03:25:36 crc kubenswrapper[4742]: I1004 03:25:36.356533 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Oct 04 03:25:36 crc kubenswrapper[4742]: I1004 03:25:36.357140 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Oct 04 03:25:36 crc kubenswrapper[4742]: I1004 03:25:36.360819 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-j6wks"] Oct 04 03:25:36 crc kubenswrapper[4742]: I1004 03:25:36.399940 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pj8n\" (UniqueName: \"kubernetes.io/projected/0778ac1c-afcb-4681-9622-29266b251908-kube-api-access-4pj8n\") pod \"keystone-db-sync-j6wks\" (UID: \"0778ac1c-afcb-4681-9622-29266b251908\") " pod="glance-kuttl-tests/keystone-db-sync-j6wks" Oct 04 03:25:36 crc kubenswrapper[4742]: I1004 03:25:36.400394 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0778ac1c-afcb-4681-9622-29266b251908-config-data\") pod \"keystone-db-sync-j6wks\" (UID: \"0778ac1c-afcb-4681-9622-29266b251908\") " pod="glance-kuttl-tests/keystone-db-sync-j6wks" Oct 04 03:25:36 crc kubenswrapper[4742]: I1004 03:25:36.501545 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pj8n\" (UniqueName: \"kubernetes.io/projected/0778ac1c-afcb-4681-9622-29266b251908-kube-api-access-4pj8n\") pod \"keystone-db-sync-j6wks\" (UID: \"0778ac1c-afcb-4681-9622-29266b251908\") " pod="glance-kuttl-tests/keystone-db-sync-j6wks" Oct 04 03:25:36 crc kubenswrapper[4742]: I1004 03:25:36.501679 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0778ac1c-afcb-4681-9622-29266b251908-config-data\") pod \"keystone-db-sync-j6wks\" (UID: \"0778ac1c-afcb-4681-9622-29266b251908\") " pod="glance-kuttl-tests/keystone-db-sync-j6wks" Oct 04 03:25:36 crc kubenswrapper[4742]: I1004 03:25:36.512210 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0778ac1c-afcb-4681-9622-29266b251908-config-data\") pod \"keystone-db-sync-j6wks\" (UID: \"0778ac1c-afcb-4681-9622-29266b251908\") " pod="glance-kuttl-tests/keystone-db-sync-j6wks" Oct 04 03:25:36 crc kubenswrapper[4742]: I1004 03:25:36.521606 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pj8n\" (UniqueName: \"kubernetes.io/projected/0778ac1c-afcb-4681-9622-29266b251908-kube-api-access-4pj8n\") pod \"keystone-db-sync-j6wks\" (UID: \"0778ac1c-afcb-4681-9622-29266b251908\") " pod="glance-kuttl-tests/keystone-db-sync-j6wks" Oct 04 03:25:36 crc kubenswrapper[4742]: I1004 03:25:36.624684 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl" event={"ID":"fda100d8-ce31-46d2-bb94-2edb9aabe58e","Type":"ContainerStarted","Data":"feedc1b345371e1dcd56129c77884eac14186fd2ea37f290d4f05bf13d1c8d88"} Oct 04 03:25:36 crc kubenswrapper[4742]: I1004 03:25:36.624744 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl" event={"ID":"fda100d8-ce31-46d2-bb94-2edb9aabe58e","Type":"ContainerStarted","Data":"36cad19fdda4b5b57afff173dd9a349dbe4b248e0801ea2fb48bf4517db97b27"} Oct 04 03:25:36 crc kubenswrapper[4742]: I1004 03:25:36.673595 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9"] Oct 04 03:25:36 crc kubenswrapper[4742]: I1004 03:25:36.675355 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9" Oct 04 03:25:36 crc kubenswrapper[4742]: I1004 03:25:36.682761 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9"] Oct 04 03:25:36 crc kubenswrapper[4742]: I1004 03:25:36.709311 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c-bundle\") pod \"c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9\" (UID: \"8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c\") " pod="openstack-operators/c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9" Oct 04 03:25:36 crc kubenswrapper[4742]: I1004 03:25:36.709856 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c-util\") pod \"c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9\" (UID: \"8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c\") " pod="openstack-operators/c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9" Oct 04 03:25:36 crc kubenswrapper[4742]: I1004 03:25:36.709895 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skkdc\" (UniqueName: \"kubernetes.io/projected/8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c-kube-api-access-skkdc\") pod \"c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9\" (UID: \"8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c\") " pod="openstack-operators/c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9" Oct 04 03:25:36 crc kubenswrapper[4742]: I1004 03:25:36.731356 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-j6wks" Oct 04 03:25:36 crc kubenswrapper[4742]: I1004 03:25:36.812038 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c-util\") pod \"c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9\" (UID: \"8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c\") " pod="openstack-operators/c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9" Oct 04 03:25:36 crc kubenswrapper[4742]: I1004 03:25:36.812153 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skkdc\" (UniqueName: \"kubernetes.io/projected/8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c-kube-api-access-skkdc\") pod \"c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9\" (UID: \"8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c\") " pod="openstack-operators/c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9" Oct 04 03:25:36 crc kubenswrapper[4742]: I1004 03:25:36.812425 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c-bundle\") pod \"c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9\" (UID: \"8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c\") " pod="openstack-operators/c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9" Oct 04 03:25:36 crc kubenswrapper[4742]: I1004 03:25:36.812666 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c-bundle\") pod \"c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9\" (UID: \"8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c\") " pod="openstack-operators/c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9" Oct 04 03:25:36 crc kubenswrapper[4742]: I1004 03:25:36.812742 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c-util\") pod \"c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9\" (UID: \"8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c\") " pod="openstack-operators/c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9" Oct 04 03:25:36 crc kubenswrapper[4742]: I1004 03:25:36.841080 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skkdc\" (UniqueName: \"kubernetes.io/projected/8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c-kube-api-access-skkdc\") pod \"c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9\" (UID: \"8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c\") " pod="openstack-operators/c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9" Oct 04 03:25:36 crc kubenswrapper[4742]: I1004 03:25:36.966374 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-j6wks"] Oct 04 03:25:36 crc kubenswrapper[4742]: W1004 03:25:36.972980 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0778ac1c_afcb_4681_9622_29266b251908.slice/crio-eb94b03fa584281ceb1817787c59cdb12a47d7098a9d564f55deeff50d71750e WatchSource:0}: Error finding container eb94b03fa584281ceb1817787c59cdb12a47d7098a9d564f55deeff50d71750e: Status 404 returned error can't find the container with id eb94b03fa584281ceb1817787c59cdb12a47d7098a9d564f55deeff50d71750e Oct 04 03:25:37 crc kubenswrapper[4742]: I1004 03:25:37.017407 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9" Oct 04 03:25:37 crc kubenswrapper[4742]: I1004 03:25:37.240593 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9"] Oct 04 03:25:37 crc kubenswrapper[4742]: I1004 03:25:37.636210 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-j6wks" event={"ID":"0778ac1c-afcb-4681-9622-29266b251908","Type":"ContainerStarted","Data":"eb94b03fa584281ceb1817787c59cdb12a47d7098a9d564f55deeff50d71750e"} Oct 04 03:25:37 crc kubenswrapper[4742]: I1004 03:25:37.640323 4742 generic.go:334] "Generic (PLEG): container finished" podID="fda100d8-ce31-46d2-bb94-2edb9aabe58e" containerID="feedc1b345371e1dcd56129c77884eac14186fd2ea37f290d4f05bf13d1c8d88" exitCode=0 Oct 04 03:25:37 crc kubenswrapper[4742]: I1004 03:25:37.640556 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl" event={"ID":"fda100d8-ce31-46d2-bb94-2edb9aabe58e","Type":"ContainerDied","Data":"feedc1b345371e1dcd56129c77884eac14186fd2ea37f290d4f05bf13d1c8d88"} Oct 04 03:25:37 crc kubenswrapper[4742]: I1004 03:25:37.644116 4742 generic.go:334] "Generic (PLEG): container finished" podID="8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c" containerID="3881eda87e0ee6cc4cafafdd668c5d1ccbb9c118b247d11cac5d02b0188bbbb5" exitCode=0 Oct 04 03:25:37 crc kubenswrapper[4742]: I1004 03:25:37.644174 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9" event={"ID":"8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c","Type":"ContainerDied","Data":"3881eda87e0ee6cc4cafafdd668c5d1ccbb9c118b247d11cac5d02b0188bbbb5"} Oct 04 03:25:37 crc kubenswrapper[4742]: I1004 03:25:37.644222 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9" event={"ID":"8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c","Type":"ContainerStarted","Data":"c67a302907f99b3d569bfe90f2cf3787e58b2dfb7100553289a36e6bc6ceebf0"} Oct 04 03:25:38 crc kubenswrapper[4742]: I1004 03:25:38.658735 4742 generic.go:334] "Generic (PLEG): container finished" podID="8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c" containerID="584252675d5d6c2850e797b9db650493c3cfa3ec52151ab2ba72a22c872b74e5" exitCode=0 Oct 04 03:25:38 crc kubenswrapper[4742]: I1004 03:25:38.658839 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9" event={"ID":"8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c","Type":"ContainerDied","Data":"584252675d5d6c2850e797b9db650493c3cfa3ec52151ab2ba72a22c872b74e5"} Oct 04 03:25:39 crc kubenswrapper[4742]: I1004 03:25:39.725470 4742 generic.go:334] "Generic (PLEG): container finished" podID="fda100d8-ce31-46d2-bb94-2edb9aabe58e" containerID="1661b7981cadfe5f8383f9cc0358d930d6f425d9a85464b5576929b24b9e5233" exitCode=0 Oct 04 03:25:39 crc kubenswrapper[4742]: I1004 03:25:39.725983 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl" event={"ID":"fda100d8-ce31-46d2-bb94-2edb9aabe58e","Type":"ContainerDied","Data":"1661b7981cadfe5f8383f9cc0358d930d6f425d9a85464b5576929b24b9e5233"} Oct 04 03:25:39 crc kubenswrapper[4742]: I1004 03:25:39.730805 4742 generic.go:334] "Generic (PLEG): container finished" podID="8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c" containerID="efc93b11fedafab7d2b2dec2bba789e13602d063d2a8a13914680d730ee48899" exitCode=0 Oct 04 03:25:39 crc kubenswrapper[4742]: I1004 03:25:39.731017 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9" event={"ID":"8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c","Type":"ContainerDied","Data":"efc93b11fedafab7d2b2dec2bba789e13602d063d2a8a13914680d730ee48899"} Oct 04 03:25:45 crc kubenswrapper[4742]: I1004 03:25:45.258865 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9" Oct 04 03:25:45 crc kubenswrapper[4742]: I1004 03:25:45.374719 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c-bundle\") pod \"8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c\" (UID: \"8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c\") " Oct 04 03:25:45 crc kubenswrapper[4742]: I1004 03:25:45.374796 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skkdc\" (UniqueName: \"kubernetes.io/projected/8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c-kube-api-access-skkdc\") pod \"8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c\" (UID: \"8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c\") " Oct 04 03:25:45 crc kubenswrapper[4742]: I1004 03:25:45.374859 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c-util\") pod \"8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c\" (UID: \"8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c\") " Oct 04 03:25:45 crc kubenswrapper[4742]: I1004 03:25:45.376328 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c-bundle" (OuterVolumeSpecName: "bundle") pod "8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c" (UID: "8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:25:45 crc kubenswrapper[4742]: I1004 03:25:45.383775 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c-kube-api-access-skkdc" (OuterVolumeSpecName: "kube-api-access-skkdc") pod "8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c" (UID: "8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c"). InnerVolumeSpecName "kube-api-access-skkdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:45 crc kubenswrapper[4742]: I1004 03:25:45.399792 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c-util" (OuterVolumeSpecName: "util") pod "8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c" (UID: "8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:25:45 crc kubenswrapper[4742]: I1004 03:25:45.477515 4742 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:45 crc kubenswrapper[4742]: I1004 03:25:45.477556 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skkdc\" (UniqueName: \"kubernetes.io/projected/8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c-kube-api-access-skkdc\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:45 crc kubenswrapper[4742]: I1004 03:25:45.477570 4742 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c-util\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:45 crc kubenswrapper[4742]: I1004 03:25:45.796500 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-j6wks" event={"ID":"0778ac1c-afcb-4681-9622-29266b251908","Type":"ContainerStarted","Data":"f9228a893222be3d29903c9a506e9fa0d7b126e0f9ee0769bf64ad01241eeff8"} Oct 04 03:25:45 crc kubenswrapper[4742]: I1004 03:25:45.806321 4742 generic.go:334] "Generic (PLEG): container finished" podID="fda100d8-ce31-46d2-bb94-2edb9aabe58e" containerID="47c7176614365910aa612d874a25ac50b197bd9375af8d1139a1cb482361515d" exitCode=0 Oct 04 03:25:45 crc kubenswrapper[4742]: I1004 03:25:45.806583 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl" event={"ID":"fda100d8-ce31-46d2-bb94-2edb9aabe58e","Type":"ContainerDied","Data":"47c7176614365910aa612d874a25ac50b197bd9375af8d1139a1cb482361515d"} Oct 04 03:25:45 crc kubenswrapper[4742]: I1004 03:25:45.811153 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9" event={"ID":"8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c","Type":"ContainerDied","Data":"c67a302907f99b3d569bfe90f2cf3787e58b2dfb7100553289a36e6bc6ceebf0"} Oct 04 03:25:45 crc kubenswrapper[4742]: I1004 03:25:45.811245 4742 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c67a302907f99b3d569bfe90f2cf3787e58b2dfb7100553289a36e6bc6ceebf0" Oct 04 03:25:45 crc kubenswrapper[4742]: I1004 03:25:45.811301 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9" Oct 04 03:25:45 crc kubenswrapper[4742]: I1004 03:25:45.829871 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-db-sync-j6wks" podStartSLOduration=1.645361552 podStartE2EDuration="9.829841645s" podCreationTimestamp="2025-10-04 03:25:36 +0000 UTC" firstStartedPulling="2025-10-04 03:25:36.976054044 +0000 UTC m=+902.908391220" lastFinishedPulling="2025-10-04 03:25:45.160534127 +0000 UTC m=+911.092871313" observedRunningTime="2025-10-04 03:25:45.821560522 +0000 UTC m=+911.753897738" watchObservedRunningTime="2025-10-04 03:25:45.829841645 +0000 UTC m=+911.762178861" Oct 04 03:25:47 crc kubenswrapper[4742]: I1004 03:25:47.140750 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl" Oct 04 03:25:47 crc kubenswrapper[4742]: I1004 03:25:47.216698 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fda100d8-ce31-46d2-bb94-2edb9aabe58e-util\") pod \"fda100d8-ce31-46d2-bb94-2edb9aabe58e\" (UID: \"fda100d8-ce31-46d2-bb94-2edb9aabe58e\") " Oct 04 03:25:47 crc kubenswrapper[4742]: I1004 03:25:47.216941 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fda100d8-ce31-46d2-bb94-2edb9aabe58e-bundle\") pod \"fda100d8-ce31-46d2-bb94-2edb9aabe58e\" (UID: \"fda100d8-ce31-46d2-bb94-2edb9aabe58e\") " Oct 04 03:25:47 crc kubenswrapper[4742]: I1004 03:25:47.217099 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ghm6\" (UniqueName: \"kubernetes.io/projected/fda100d8-ce31-46d2-bb94-2edb9aabe58e-kube-api-access-7ghm6\") pod \"fda100d8-ce31-46d2-bb94-2edb9aabe58e\" (UID: \"fda100d8-ce31-46d2-bb94-2edb9aabe58e\") " Oct 04 03:25:47 crc kubenswrapper[4742]: I1004 03:25:47.217770 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fda100d8-ce31-46d2-bb94-2edb9aabe58e-bundle" (OuterVolumeSpecName: "bundle") pod "fda100d8-ce31-46d2-bb94-2edb9aabe58e" (UID: "fda100d8-ce31-46d2-bb94-2edb9aabe58e"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:25:47 crc kubenswrapper[4742]: I1004 03:25:47.218021 4742 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fda100d8-ce31-46d2-bb94-2edb9aabe58e-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:47 crc kubenswrapper[4742]: I1004 03:25:47.231458 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda100d8-ce31-46d2-bb94-2edb9aabe58e-kube-api-access-7ghm6" (OuterVolumeSpecName: "kube-api-access-7ghm6") pod "fda100d8-ce31-46d2-bb94-2edb9aabe58e" (UID: "fda100d8-ce31-46d2-bb94-2edb9aabe58e"). InnerVolumeSpecName "kube-api-access-7ghm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:47 crc kubenswrapper[4742]: I1004 03:25:47.234011 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fda100d8-ce31-46d2-bb94-2edb9aabe58e-util" (OuterVolumeSpecName: "util") pod "fda100d8-ce31-46d2-bb94-2edb9aabe58e" (UID: "fda100d8-ce31-46d2-bb94-2edb9aabe58e"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:25:47 crc kubenswrapper[4742]: I1004 03:25:47.318997 4742 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fda100d8-ce31-46d2-bb94-2edb9aabe58e-util\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:47 crc kubenswrapper[4742]: I1004 03:25:47.319035 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ghm6\" (UniqueName: \"kubernetes.io/projected/fda100d8-ce31-46d2-bb94-2edb9aabe58e-kube-api-access-7ghm6\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:47 crc kubenswrapper[4742]: I1004 03:25:47.830487 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl" event={"ID":"fda100d8-ce31-46d2-bb94-2edb9aabe58e","Type":"ContainerDied","Data":"36cad19fdda4b5b57afff173dd9a349dbe4b248e0801ea2fb48bf4517db97b27"} Oct 04 03:25:47 crc kubenswrapper[4742]: I1004 03:25:47.830555 4742 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="36cad19fdda4b5b57afff173dd9a349dbe4b248e0801ea2fb48bf4517db97b27" Oct 04 03:25:47 crc kubenswrapper[4742]: I1004 03:25:47.830562 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl" Oct 04 03:25:48 crc kubenswrapper[4742]: I1004 03:25:48.842105 4742 generic.go:334] "Generic (PLEG): container finished" podID="0778ac1c-afcb-4681-9622-29266b251908" containerID="f9228a893222be3d29903c9a506e9fa0d7b126e0f9ee0769bf64ad01241eeff8" exitCode=0 Oct 04 03:25:48 crc kubenswrapper[4742]: I1004 03:25:48.842225 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-j6wks" event={"ID":"0778ac1c-afcb-4681-9622-29266b251908","Type":"ContainerDied","Data":"f9228a893222be3d29903c9a506e9fa0d7b126e0f9ee0769bf64ad01241eeff8"} Oct 04 03:25:50 crc kubenswrapper[4742]: I1004 03:25:50.198100 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-j6wks" Oct 04 03:25:50 crc kubenswrapper[4742]: I1004 03:25:50.280379 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pj8n\" (UniqueName: \"kubernetes.io/projected/0778ac1c-afcb-4681-9622-29266b251908-kube-api-access-4pj8n\") pod \"0778ac1c-afcb-4681-9622-29266b251908\" (UID: \"0778ac1c-afcb-4681-9622-29266b251908\") " Oct 04 03:25:50 crc kubenswrapper[4742]: I1004 03:25:50.280855 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0778ac1c-afcb-4681-9622-29266b251908-config-data\") pod \"0778ac1c-afcb-4681-9622-29266b251908\" (UID: \"0778ac1c-afcb-4681-9622-29266b251908\") " Oct 04 03:25:50 crc kubenswrapper[4742]: I1004 03:25:50.289450 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0778ac1c-afcb-4681-9622-29266b251908-kube-api-access-4pj8n" (OuterVolumeSpecName: "kube-api-access-4pj8n") pod "0778ac1c-afcb-4681-9622-29266b251908" (UID: "0778ac1c-afcb-4681-9622-29266b251908"). InnerVolumeSpecName "kube-api-access-4pj8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:50 crc kubenswrapper[4742]: I1004 03:25:50.331620 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0778ac1c-afcb-4681-9622-29266b251908-config-data" (OuterVolumeSpecName: "config-data") pod "0778ac1c-afcb-4681-9622-29266b251908" (UID: "0778ac1c-afcb-4681-9622-29266b251908"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:50 crc kubenswrapper[4742]: I1004 03:25:50.382348 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pj8n\" (UniqueName: \"kubernetes.io/projected/0778ac1c-afcb-4681-9622-29266b251908-kube-api-access-4pj8n\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:50 crc kubenswrapper[4742]: I1004 03:25:50.382406 4742 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0778ac1c-afcb-4681-9622-29266b251908-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:50 crc kubenswrapper[4742]: I1004 03:25:50.856694 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-j6wks" event={"ID":"0778ac1c-afcb-4681-9622-29266b251908","Type":"ContainerDied","Data":"eb94b03fa584281ceb1817787c59cdb12a47d7098a9d564f55deeff50d71750e"} Oct 04 03:25:50 crc kubenswrapper[4742]: I1004 03:25:50.856737 4742 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb94b03fa584281ceb1817787c59cdb12a47d7098a9d564f55deeff50d71750e" Oct 04 03:25:50 crc kubenswrapper[4742]: I1004 03:25:50.856778 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-j6wks" Oct 04 03:25:51 crc kubenswrapper[4742]: I1004 03:25:51.095745 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-4rtpb"] Oct 04 03:25:51 crc kubenswrapper[4742]: E1004 03:25:51.096003 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fda100d8-ce31-46d2-bb94-2edb9aabe58e" containerName="extract" Oct 04 03:25:51 crc kubenswrapper[4742]: I1004 03:25:51.096015 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="fda100d8-ce31-46d2-bb94-2edb9aabe58e" containerName="extract" Oct 04 03:25:51 crc kubenswrapper[4742]: E1004 03:25:51.096028 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fda100d8-ce31-46d2-bb94-2edb9aabe58e" containerName="pull" Oct 04 03:25:51 crc kubenswrapper[4742]: I1004 03:25:51.096035 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="fda100d8-ce31-46d2-bb94-2edb9aabe58e" containerName="pull" Oct 04 03:25:51 crc kubenswrapper[4742]: E1004 03:25:51.096045 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c" containerName="util" Oct 04 03:25:51 crc kubenswrapper[4742]: I1004 03:25:51.096051 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c" containerName="util" Oct 04 03:25:51 crc kubenswrapper[4742]: E1004 03:25:51.096064 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c" containerName="extract" Oct 04 03:25:51 crc kubenswrapper[4742]: I1004 03:25:51.096069 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c" containerName="extract" Oct 04 03:25:51 crc kubenswrapper[4742]: E1004 03:25:51.096078 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fda100d8-ce31-46d2-bb94-2edb9aabe58e" containerName="util" Oct 04 03:25:51 crc kubenswrapper[4742]: I1004 03:25:51.096084 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="fda100d8-ce31-46d2-bb94-2edb9aabe58e" containerName="util" Oct 04 03:25:51 crc kubenswrapper[4742]: E1004 03:25:51.096100 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c" containerName="pull" Oct 04 03:25:51 crc kubenswrapper[4742]: I1004 03:25:51.096107 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c" containerName="pull" Oct 04 03:25:51 crc kubenswrapper[4742]: E1004 03:25:51.096115 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0778ac1c-afcb-4681-9622-29266b251908" containerName="keystone-db-sync" Oct 04 03:25:51 crc kubenswrapper[4742]: I1004 03:25:51.096121 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="0778ac1c-afcb-4681-9622-29266b251908" containerName="keystone-db-sync" Oct 04 03:25:51 crc kubenswrapper[4742]: I1004 03:25:51.096215 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="fda100d8-ce31-46d2-bb94-2edb9aabe58e" containerName="extract" Oct 04 03:25:51 crc kubenswrapper[4742]: I1004 03:25:51.096228 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c" containerName="extract" Oct 04 03:25:51 crc kubenswrapper[4742]: I1004 03:25:51.096240 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="0778ac1c-afcb-4681-9622-29266b251908" containerName="keystone-db-sync" Oct 04 03:25:51 crc kubenswrapper[4742]: I1004 03:25:51.096686 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-4rtpb" Oct 04 03:25:51 crc kubenswrapper[4742]: I1004 03:25:51.097926 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-whckd" Oct 04 03:25:51 crc kubenswrapper[4742]: I1004 03:25:51.099300 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Oct 04 03:25:51 crc kubenswrapper[4742]: I1004 03:25:51.099305 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Oct 04 03:25:51 crc kubenswrapper[4742]: I1004 03:25:51.099363 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Oct 04 03:25:51 crc kubenswrapper[4742]: I1004 03:25:51.106337 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-4rtpb"] Oct 04 03:25:51 crc kubenswrapper[4742]: I1004 03:25:51.193351 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1593a5c-e6be-4abb-9d84-c78c90c8dcf6-config-data\") pod \"keystone-bootstrap-4rtpb\" (UID: \"d1593a5c-e6be-4abb-9d84-c78c90c8dcf6\") " pod="glance-kuttl-tests/keystone-bootstrap-4rtpb" Oct 04 03:25:51 crc kubenswrapper[4742]: I1004 03:25:51.193432 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d1593a5c-e6be-4abb-9d84-c78c90c8dcf6-fernet-keys\") pod \"keystone-bootstrap-4rtpb\" (UID: \"d1593a5c-e6be-4abb-9d84-c78c90c8dcf6\") " pod="glance-kuttl-tests/keystone-bootstrap-4rtpb" Oct 04 03:25:51 crc kubenswrapper[4742]: I1004 03:25:51.193452 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1593a5c-e6be-4abb-9d84-c78c90c8dcf6-scripts\") pod \"keystone-bootstrap-4rtpb\" (UID: \"d1593a5c-e6be-4abb-9d84-c78c90c8dcf6\") " pod="glance-kuttl-tests/keystone-bootstrap-4rtpb" Oct 04 03:25:51 crc kubenswrapper[4742]: I1004 03:25:51.193476 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d1593a5c-e6be-4abb-9d84-c78c90c8dcf6-credential-keys\") pod \"keystone-bootstrap-4rtpb\" (UID: \"d1593a5c-e6be-4abb-9d84-c78c90c8dcf6\") " pod="glance-kuttl-tests/keystone-bootstrap-4rtpb" Oct 04 03:25:51 crc kubenswrapper[4742]: I1004 03:25:51.193503 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74tpd\" (UniqueName: \"kubernetes.io/projected/d1593a5c-e6be-4abb-9d84-c78c90c8dcf6-kube-api-access-74tpd\") pod \"keystone-bootstrap-4rtpb\" (UID: \"d1593a5c-e6be-4abb-9d84-c78c90c8dcf6\") " pod="glance-kuttl-tests/keystone-bootstrap-4rtpb" Oct 04 03:25:51 crc kubenswrapper[4742]: I1004 03:25:51.295148 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1593a5c-e6be-4abb-9d84-c78c90c8dcf6-config-data\") pod \"keystone-bootstrap-4rtpb\" (UID: \"d1593a5c-e6be-4abb-9d84-c78c90c8dcf6\") " pod="glance-kuttl-tests/keystone-bootstrap-4rtpb" Oct 04 03:25:51 crc kubenswrapper[4742]: I1004 03:25:51.295227 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d1593a5c-e6be-4abb-9d84-c78c90c8dcf6-fernet-keys\") pod \"keystone-bootstrap-4rtpb\" (UID: \"d1593a5c-e6be-4abb-9d84-c78c90c8dcf6\") " pod="glance-kuttl-tests/keystone-bootstrap-4rtpb" Oct 04 03:25:51 crc kubenswrapper[4742]: I1004 03:25:51.295245 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1593a5c-e6be-4abb-9d84-c78c90c8dcf6-scripts\") pod \"keystone-bootstrap-4rtpb\" (UID: \"d1593a5c-e6be-4abb-9d84-c78c90c8dcf6\") " pod="glance-kuttl-tests/keystone-bootstrap-4rtpb" Oct 04 03:25:51 crc kubenswrapper[4742]: I1004 03:25:51.295289 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d1593a5c-e6be-4abb-9d84-c78c90c8dcf6-credential-keys\") pod \"keystone-bootstrap-4rtpb\" (UID: \"d1593a5c-e6be-4abb-9d84-c78c90c8dcf6\") " pod="glance-kuttl-tests/keystone-bootstrap-4rtpb" Oct 04 03:25:51 crc kubenswrapper[4742]: I1004 03:25:51.295319 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74tpd\" (UniqueName: \"kubernetes.io/projected/d1593a5c-e6be-4abb-9d84-c78c90c8dcf6-kube-api-access-74tpd\") pod \"keystone-bootstrap-4rtpb\" (UID: \"d1593a5c-e6be-4abb-9d84-c78c90c8dcf6\") " pod="glance-kuttl-tests/keystone-bootstrap-4rtpb" Oct 04 03:25:51 crc kubenswrapper[4742]: I1004 03:25:51.300819 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1593a5c-e6be-4abb-9d84-c78c90c8dcf6-scripts\") pod \"keystone-bootstrap-4rtpb\" (UID: \"d1593a5c-e6be-4abb-9d84-c78c90c8dcf6\") " pod="glance-kuttl-tests/keystone-bootstrap-4rtpb" Oct 04 03:25:51 crc kubenswrapper[4742]: I1004 03:25:51.301251 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d1593a5c-e6be-4abb-9d84-c78c90c8dcf6-credential-keys\") pod \"keystone-bootstrap-4rtpb\" (UID: \"d1593a5c-e6be-4abb-9d84-c78c90c8dcf6\") " pod="glance-kuttl-tests/keystone-bootstrap-4rtpb" Oct 04 03:25:51 crc kubenswrapper[4742]: I1004 03:25:51.301873 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d1593a5c-e6be-4abb-9d84-c78c90c8dcf6-fernet-keys\") pod \"keystone-bootstrap-4rtpb\" (UID: \"d1593a5c-e6be-4abb-9d84-c78c90c8dcf6\") " pod="glance-kuttl-tests/keystone-bootstrap-4rtpb" Oct 04 03:25:51 crc kubenswrapper[4742]: I1004 03:25:51.306434 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1593a5c-e6be-4abb-9d84-c78c90c8dcf6-config-data\") pod \"keystone-bootstrap-4rtpb\" (UID: \"d1593a5c-e6be-4abb-9d84-c78c90c8dcf6\") " pod="glance-kuttl-tests/keystone-bootstrap-4rtpb" Oct 04 03:25:51 crc kubenswrapper[4742]: I1004 03:25:51.323951 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74tpd\" (UniqueName: \"kubernetes.io/projected/d1593a5c-e6be-4abb-9d84-c78c90c8dcf6-kube-api-access-74tpd\") pod \"keystone-bootstrap-4rtpb\" (UID: \"d1593a5c-e6be-4abb-9d84-c78c90c8dcf6\") " pod="glance-kuttl-tests/keystone-bootstrap-4rtpb" Oct 04 03:25:51 crc kubenswrapper[4742]: I1004 03:25:51.412070 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-4rtpb" Oct 04 03:25:51 crc kubenswrapper[4742]: I1004 03:25:51.667734 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-4rtpb"] Oct 04 03:25:51 crc kubenswrapper[4742]: W1004 03:25:51.679122 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1593a5c_e6be_4abb_9d84_c78c90c8dcf6.slice/crio-7748a95824d659d11a1e6565362b3b95501846b3859c04d449d0fb1c7e8ec31b WatchSource:0}: Error finding container 7748a95824d659d11a1e6565362b3b95501846b3859c04d449d0fb1c7e8ec31b: Status 404 returned error can't find the container with id 7748a95824d659d11a1e6565362b3b95501846b3859c04d449d0fb1c7e8ec31b Oct 04 03:25:51 crc kubenswrapper[4742]: I1004 03:25:51.868409 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-4rtpb" event={"ID":"d1593a5c-e6be-4abb-9d84-c78c90c8dcf6","Type":"ContainerStarted","Data":"7748a95824d659d11a1e6565362b3b95501846b3859c04d449d0fb1c7e8ec31b"} Oct 04 03:25:52 crc kubenswrapper[4742]: I1004 03:25:52.876330 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-4rtpb" event={"ID":"d1593a5c-e6be-4abb-9d84-c78c90c8dcf6","Type":"ContainerStarted","Data":"b8cabab5b57e3720fdf1f8871b6ba110850453e8d3e7f0e7823b70a8470e6366"} Oct 04 03:25:52 crc kubenswrapper[4742]: I1004 03:25:52.901843 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-bootstrap-4rtpb" podStartSLOduration=1.901818907 podStartE2EDuration="1.901818907s" podCreationTimestamp="2025-10-04 03:25:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:25:52.899423285 +0000 UTC m=+918.831760511" watchObservedRunningTime="2025-10-04 03:25:52.901818907 +0000 UTC m=+918.834156093" Oct 04 03:25:54 crc kubenswrapper[4742]: I1004 03:25:54.781738 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f47cfc687-xl4hq"] Oct 04 03:25:54 crc kubenswrapper[4742]: I1004 03:25:54.784887 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f47cfc687-xl4hq" Oct 04 03:25:54 crc kubenswrapper[4742]: I1004 03:25:54.788846 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-service-cert" Oct 04 03:25:54 crc kubenswrapper[4742]: I1004 03:25:54.796827 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-lgct4" Oct 04 03:25:54 crc kubenswrapper[4742]: I1004 03:25:54.812542 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f47cfc687-xl4hq"] Oct 04 03:25:54 crc kubenswrapper[4742]: I1004 03:25:54.889628 4742 generic.go:334] "Generic (PLEG): container finished" podID="d1593a5c-e6be-4abb-9d84-c78c90c8dcf6" containerID="b8cabab5b57e3720fdf1f8871b6ba110850453e8d3e7f0e7823b70a8470e6366" exitCode=0 Oct 04 03:25:54 crc kubenswrapper[4742]: I1004 03:25:54.889683 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-4rtpb" event={"ID":"d1593a5c-e6be-4abb-9d84-c78c90c8dcf6","Type":"ContainerDied","Data":"b8cabab5b57e3720fdf1f8871b6ba110850453e8d3e7f0e7823b70a8470e6366"} Oct 04 03:25:54 crc kubenswrapper[4742]: I1004 03:25:54.957772 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/744a78c0-6d93-4a39-bb0c-68f80ff9716b-webhook-cert\") pod \"swift-operator-controller-manager-5f47cfc687-xl4hq\" (UID: \"744a78c0-6d93-4a39-bb0c-68f80ff9716b\") " pod="openstack-operators/swift-operator-controller-manager-5f47cfc687-xl4hq" Oct 04 03:25:54 crc kubenswrapper[4742]: I1004 03:25:54.957821 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/744a78c0-6d93-4a39-bb0c-68f80ff9716b-apiservice-cert\") pod \"swift-operator-controller-manager-5f47cfc687-xl4hq\" (UID: \"744a78c0-6d93-4a39-bb0c-68f80ff9716b\") " pod="openstack-operators/swift-operator-controller-manager-5f47cfc687-xl4hq" Oct 04 03:25:54 crc kubenswrapper[4742]: I1004 03:25:54.957843 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7v7ng\" (UniqueName: \"kubernetes.io/projected/744a78c0-6d93-4a39-bb0c-68f80ff9716b-kube-api-access-7v7ng\") pod \"swift-operator-controller-manager-5f47cfc687-xl4hq\" (UID: \"744a78c0-6d93-4a39-bb0c-68f80ff9716b\") " pod="openstack-operators/swift-operator-controller-manager-5f47cfc687-xl4hq" Oct 04 03:25:55 crc kubenswrapper[4742]: I1004 03:25:55.059653 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/744a78c0-6d93-4a39-bb0c-68f80ff9716b-webhook-cert\") pod \"swift-operator-controller-manager-5f47cfc687-xl4hq\" (UID: \"744a78c0-6d93-4a39-bb0c-68f80ff9716b\") " pod="openstack-operators/swift-operator-controller-manager-5f47cfc687-xl4hq" Oct 04 03:25:55 crc kubenswrapper[4742]: I1004 03:25:55.059715 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/744a78c0-6d93-4a39-bb0c-68f80ff9716b-apiservice-cert\") pod \"swift-operator-controller-manager-5f47cfc687-xl4hq\" (UID: \"744a78c0-6d93-4a39-bb0c-68f80ff9716b\") " pod="openstack-operators/swift-operator-controller-manager-5f47cfc687-xl4hq" Oct 04 03:25:55 crc kubenswrapper[4742]: I1004 03:25:55.059742 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7v7ng\" (UniqueName: \"kubernetes.io/projected/744a78c0-6d93-4a39-bb0c-68f80ff9716b-kube-api-access-7v7ng\") pod \"swift-operator-controller-manager-5f47cfc687-xl4hq\" (UID: \"744a78c0-6d93-4a39-bb0c-68f80ff9716b\") " pod="openstack-operators/swift-operator-controller-manager-5f47cfc687-xl4hq" Oct 04 03:25:55 crc kubenswrapper[4742]: I1004 03:25:55.068605 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/744a78c0-6d93-4a39-bb0c-68f80ff9716b-apiservice-cert\") pod \"swift-operator-controller-manager-5f47cfc687-xl4hq\" (UID: \"744a78c0-6d93-4a39-bb0c-68f80ff9716b\") " pod="openstack-operators/swift-operator-controller-manager-5f47cfc687-xl4hq" Oct 04 03:25:55 crc kubenswrapper[4742]: I1004 03:25:55.068754 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/744a78c0-6d93-4a39-bb0c-68f80ff9716b-webhook-cert\") pod \"swift-operator-controller-manager-5f47cfc687-xl4hq\" (UID: \"744a78c0-6d93-4a39-bb0c-68f80ff9716b\") " pod="openstack-operators/swift-operator-controller-manager-5f47cfc687-xl4hq" Oct 04 03:25:55 crc kubenswrapper[4742]: I1004 03:25:55.080580 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7v7ng\" (UniqueName: \"kubernetes.io/projected/744a78c0-6d93-4a39-bb0c-68f80ff9716b-kube-api-access-7v7ng\") pod \"swift-operator-controller-manager-5f47cfc687-xl4hq\" (UID: \"744a78c0-6d93-4a39-bb0c-68f80ff9716b\") " pod="openstack-operators/swift-operator-controller-manager-5f47cfc687-xl4hq" Oct 04 03:25:55 crc kubenswrapper[4742]: I1004 03:25:55.126784 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f47cfc687-xl4hq" Oct 04 03:25:55 crc kubenswrapper[4742]: I1004 03:25:55.642871 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f47cfc687-xl4hq"] Oct 04 03:25:55 crc kubenswrapper[4742]: W1004 03:25:55.676951 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod744a78c0_6d93_4a39_bb0c_68f80ff9716b.slice/crio-ab8e717dff27513a0dbd15cd0d004ae94c839accaa062e2e93c432d022eaeca5 WatchSource:0}: Error finding container ab8e717dff27513a0dbd15cd0d004ae94c839accaa062e2e93c432d022eaeca5: Status 404 returned error can't find the container with id ab8e717dff27513a0dbd15cd0d004ae94c839accaa062e2e93c432d022eaeca5 Oct 04 03:25:55 crc kubenswrapper[4742]: I1004 03:25:55.680465 4742 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 03:25:55 crc kubenswrapper[4742]: I1004 03:25:55.897328 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f47cfc687-xl4hq" event={"ID":"744a78c0-6d93-4a39-bb0c-68f80ff9716b","Type":"ContainerStarted","Data":"ab8e717dff27513a0dbd15cd0d004ae94c839accaa062e2e93c432d022eaeca5"} Oct 04 03:25:56 crc kubenswrapper[4742]: I1004 03:25:56.212442 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-4rtpb" Oct 04 03:25:56 crc kubenswrapper[4742]: I1004 03:25:56.309838 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d1593a5c-e6be-4abb-9d84-c78c90c8dcf6-fernet-keys\") pod \"d1593a5c-e6be-4abb-9d84-c78c90c8dcf6\" (UID: \"d1593a5c-e6be-4abb-9d84-c78c90c8dcf6\") " Oct 04 03:25:56 crc kubenswrapper[4742]: I1004 03:25:56.309886 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1593a5c-e6be-4abb-9d84-c78c90c8dcf6-config-data\") pod \"d1593a5c-e6be-4abb-9d84-c78c90c8dcf6\" (UID: \"d1593a5c-e6be-4abb-9d84-c78c90c8dcf6\") " Oct 04 03:25:56 crc kubenswrapper[4742]: I1004 03:25:56.309906 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d1593a5c-e6be-4abb-9d84-c78c90c8dcf6-credential-keys\") pod \"d1593a5c-e6be-4abb-9d84-c78c90c8dcf6\" (UID: \"d1593a5c-e6be-4abb-9d84-c78c90c8dcf6\") " Oct 04 03:25:56 crc kubenswrapper[4742]: I1004 03:25:56.309928 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1593a5c-e6be-4abb-9d84-c78c90c8dcf6-scripts\") pod \"d1593a5c-e6be-4abb-9d84-c78c90c8dcf6\" (UID: \"d1593a5c-e6be-4abb-9d84-c78c90c8dcf6\") " Oct 04 03:25:56 crc kubenswrapper[4742]: I1004 03:25:56.309969 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74tpd\" (UniqueName: \"kubernetes.io/projected/d1593a5c-e6be-4abb-9d84-c78c90c8dcf6-kube-api-access-74tpd\") pod \"d1593a5c-e6be-4abb-9d84-c78c90c8dcf6\" (UID: \"d1593a5c-e6be-4abb-9d84-c78c90c8dcf6\") " Oct 04 03:25:56 crc kubenswrapper[4742]: I1004 03:25:56.315855 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1593a5c-e6be-4abb-9d84-c78c90c8dcf6-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "d1593a5c-e6be-4abb-9d84-c78c90c8dcf6" (UID: "d1593a5c-e6be-4abb-9d84-c78c90c8dcf6"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:56 crc kubenswrapper[4742]: I1004 03:25:56.335693 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1593a5c-e6be-4abb-9d84-c78c90c8dcf6-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "d1593a5c-e6be-4abb-9d84-c78c90c8dcf6" (UID: "d1593a5c-e6be-4abb-9d84-c78c90c8dcf6"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:56 crc kubenswrapper[4742]: I1004 03:25:56.336447 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1593a5c-e6be-4abb-9d84-c78c90c8dcf6-kube-api-access-74tpd" (OuterVolumeSpecName: "kube-api-access-74tpd") pod "d1593a5c-e6be-4abb-9d84-c78c90c8dcf6" (UID: "d1593a5c-e6be-4abb-9d84-c78c90c8dcf6"). InnerVolumeSpecName "kube-api-access-74tpd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:25:56 crc kubenswrapper[4742]: I1004 03:25:56.337437 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1593a5c-e6be-4abb-9d84-c78c90c8dcf6-scripts" (OuterVolumeSpecName: "scripts") pod "d1593a5c-e6be-4abb-9d84-c78c90c8dcf6" (UID: "d1593a5c-e6be-4abb-9d84-c78c90c8dcf6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:56 crc kubenswrapper[4742]: I1004 03:25:56.344091 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1593a5c-e6be-4abb-9d84-c78c90c8dcf6-config-data" (OuterVolumeSpecName: "config-data") pod "d1593a5c-e6be-4abb-9d84-c78c90c8dcf6" (UID: "d1593a5c-e6be-4abb-9d84-c78c90c8dcf6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:25:56 crc kubenswrapper[4742]: I1004 03:25:56.411692 4742 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d1593a5c-e6be-4abb-9d84-c78c90c8dcf6-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:56 crc kubenswrapper[4742]: I1004 03:25:56.411801 4742 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1593a5c-e6be-4abb-9d84-c78c90c8dcf6-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:56 crc kubenswrapper[4742]: I1004 03:25:56.411813 4742 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d1593a5c-e6be-4abb-9d84-c78c90c8dcf6-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:56 crc kubenswrapper[4742]: I1004 03:25:56.411825 4742 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1593a5c-e6be-4abb-9d84-c78c90c8dcf6-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:56 crc kubenswrapper[4742]: I1004 03:25:56.411837 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74tpd\" (UniqueName: \"kubernetes.io/projected/d1593a5c-e6be-4abb-9d84-c78c90c8dcf6-kube-api-access-74tpd\") on node \"crc\" DevicePath \"\"" Oct 04 03:25:56 crc kubenswrapper[4742]: I1004 03:25:56.910762 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-4rtpb" event={"ID":"d1593a5c-e6be-4abb-9d84-c78c90c8dcf6","Type":"ContainerDied","Data":"7748a95824d659d11a1e6565362b3b95501846b3859c04d449d0fb1c7e8ec31b"} Oct 04 03:25:56 crc kubenswrapper[4742]: I1004 03:25:56.910807 4742 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7748a95824d659d11a1e6565362b3b95501846b3859c04d449d0fb1c7e8ec31b" Oct 04 03:25:56 crc kubenswrapper[4742]: I1004 03:25:56.910805 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-4rtpb" Oct 04 03:25:57 crc kubenswrapper[4742]: I1004 03:25:57.326924 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-5b74dd9df7-n4fhz"] Oct 04 03:25:57 crc kubenswrapper[4742]: E1004 03:25:57.327341 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1593a5c-e6be-4abb-9d84-c78c90c8dcf6" containerName="keystone-bootstrap" Oct 04 03:25:57 crc kubenswrapper[4742]: I1004 03:25:57.327363 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1593a5c-e6be-4abb-9d84-c78c90c8dcf6" containerName="keystone-bootstrap" Oct 04 03:25:57 crc kubenswrapper[4742]: I1004 03:25:57.327560 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1593a5c-e6be-4abb-9d84-c78c90c8dcf6" containerName="keystone-bootstrap" Oct 04 03:25:57 crc kubenswrapper[4742]: I1004 03:25:57.328069 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-5b74dd9df7-n4fhz" Oct 04 03:25:57 crc kubenswrapper[4742]: I1004 03:25:57.331606 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Oct 04 03:25:57 crc kubenswrapper[4742]: I1004 03:25:57.331834 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-whckd" Oct 04 03:25:57 crc kubenswrapper[4742]: I1004 03:25:57.332049 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Oct 04 03:25:57 crc kubenswrapper[4742]: I1004 03:25:57.332203 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Oct 04 03:25:57 crc kubenswrapper[4742]: I1004 03:25:57.348170 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-5b74dd9df7-n4fhz"] Oct 04 03:25:57 crc kubenswrapper[4742]: I1004 03:25:57.529890 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f99f948a-5fbd-4292-abcd-c8c64992634e-credential-keys\") pod \"keystone-5b74dd9df7-n4fhz\" (UID: \"f99f948a-5fbd-4292-abcd-c8c64992634e\") " pod="glance-kuttl-tests/keystone-5b74dd9df7-n4fhz" Oct 04 03:25:57 crc kubenswrapper[4742]: I1004 03:25:57.529959 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f99f948a-5fbd-4292-abcd-c8c64992634e-config-data\") pod \"keystone-5b74dd9df7-n4fhz\" (UID: \"f99f948a-5fbd-4292-abcd-c8c64992634e\") " pod="glance-kuttl-tests/keystone-5b74dd9df7-n4fhz" Oct 04 03:25:57 crc kubenswrapper[4742]: I1004 03:25:57.530199 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f99f948a-5fbd-4292-abcd-c8c64992634e-scripts\") pod \"keystone-5b74dd9df7-n4fhz\" (UID: \"f99f948a-5fbd-4292-abcd-c8c64992634e\") " pod="glance-kuttl-tests/keystone-5b74dd9df7-n4fhz" Oct 04 03:25:57 crc kubenswrapper[4742]: I1004 03:25:57.530256 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5k5kk\" (UniqueName: \"kubernetes.io/projected/f99f948a-5fbd-4292-abcd-c8c64992634e-kube-api-access-5k5kk\") pod \"keystone-5b74dd9df7-n4fhz\" (UID: \"f99f948a-5fbd-4292-abcd-c8c64992634e\") " pod="glance-kuttl-tests/keystone-5b74dd9df7-n4fhz" Oct 04 03:25:57 crc kubenswrapper[4742]: I1004 03:25:57.530438 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f99f948a-5fbd-4292-abcd-c8c64992634e-fernet-keys\") pod \"keystone-5b74dd9df7-n4fhz\" (UID: \"f99f948a-5fbd-4292-abcd-c8c64992634e\") " pod="glance-kuttl-tests/keystone-5b74dd9df7-n4fhz" Oct 04 03:25:57 crc kubenswrapper[4742]: I1004 03:25:57.632110 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f99f948a-5fbd-4292-abcd-c8c64992634e-scripts\") pod \"keystone-5b74dd9df7-n4fhz\" (UID: \"f99f948a-5fbd-4292-abcd-c8c64992634e\") " pod="glance-kuttl-tests/keystone-5b74dd9df7-n4fhz" Oct 04 03:25:57 crc kubenswrapper[4742]: I1004 03:25:57.632169 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5k5kk\" (UniqueName: \"kubernetes.io/projected/f99f948a-5fbd-4292-abcd-c8c64992634e-kube-api-access-5k5kk\") pod \"keystone-5b74dd9df7-n4fhz\" (UID: \"f99f948a-5fbd-4292-abcd-c8c64992634e\") " pod="glance-kuttl-tests/keystone-5b74dd9df7-n4fhz" Oct 04 03:25:57 crc kubenswrapper[4742]: I1004 03:25:57.632230 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f99f948a-5fbd-4292-abcd-c8c64992634e-fernet-keys\") pod \"keystone-5b74dd9df7-n4fhz\" (UID: \"f99f948a-5fbd-4292-abcd-c8c64992634e\") " pod="glance-kuttl-tests/keystone-5b74dd9df7-n4fhz" Oct 04 03:25:57 crc kubenswrapper[4742]: I1004 03:25:57.632264 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f99f948a-5fbd-4292-abcd-c8c64992634e-credential-keys\") pod \"keystone-5b74dd9df7-n4fhz\" (UID: \"f99f948a-5fbd-4292-abcd-c8c64992634e\") " pod="glance-kuttl-tests/keystone-5b74dd9df7-n4fhz" Oct 04 03:25:57 crc kubenswrapper[4742]: I1004 03:25:57.632312 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f99f948a-5fbd-4292-abcd-c8c64992634e-config-data\") pod \"keystone-5b74dd9df7-n4fhz\" (UID: \"f99f948a-5fbd-4292-abcd-c8c64992634e\") " pod="glance-kuttl-tests/keystone-5b74dd9df7-n4fhz" Oct 04 03:25:57 crc kubenswrapper[4742]: I1004 03:25:57.641060 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f99f948a-5fbd-4292-abcd-c8c64992634e-fernet-keys\") pod \"keystone-5b74dd9df7-n4fhz\" (UID: \"f99f948a-5fbd-4292-abcd-c8c64992634e\") " pod="glance-kuttl-tests/keystone-5b74dd9df7-n4fhz" Oct 04 03:25:57 crc kubenswrapper[4742]: I1004 03:25:57.646125 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f99f948a-5fbd-4292-abcd-c8c64992634e-config-data\") pod \"keystone-5b74dd9df7-n4fhz\" (UID: \"f99f948a-5fbd-4292-abcd-c8c64992634e\") " pod="glance-kuttl-tests/keystone-5b74dd9df7-n4fhz" Oct 04 03:25:57 crc kubenswrapper[4742]: I1004 03:25:57.648734 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f99f948a-5fbd-4292-abcd-c8c64992634e-scripts\") pod \"keystone-5b74dd9df7-n4fhz\" (UID: \"f99f948a-5fbd-4292-abcd-c8c64992634e\") " pod="glance-kuttl-tests/keystone-5b74dd9df7-n4fhz" Oct 04 03:25:57 crc kubenswrapper[4742]: I1004 03:25:57.652398 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5k5kk\" (UniqueName: \"kubernetes.io/projected/f99f948a-5fbd-4292-abcd-c8c64992634e-kube-api-access-5k5kk\") pod \"keystone-5b74dd9df7-n4fhz\" (UID: \"f99f948a-5fbd-4292-abcd-c8c64992634e\") " pod="glance-kuttl-tests/keystone-5b74dd9df7-n4fhz" Oct 04 03:25:57 crc kubenswrapper[4742]: I1004 03:25:57.663928 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f99f948a-5fbd-4292-abcd-c8c64992634e-credential-keys\") pod \"keystone-5b74dd9df7-n4fhz\" (UID: \"f99f948a-5fbd-4292-abcd-c8c64992634e\") " pod="glance-kuttl-tests/keystone-5b74dd9df7-n4fhz" Oct 04 03:25:57 crc kubenswrapper[4742]: I1004 03:25:57.921461 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f47cfc687-xl4hq" event={"ID":"744a78c0-6d93-4a39-bb0c-68f80ff9716b","Type":"ContainerStarted","Data":"76a51b32d6fb8309ed2b0695869a1f88b3471aaa47ce87da861fc3e34ccd5192"} Oct 04 03:25:57 crc kubenswrapper[4742]: I1004 03:25:57.953163 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-5b74dd9df7-n4fhz" Oct 04 03:25:58 crc kubenswrapper[4742]: I1004 03:25:58.524369 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-5b74dd9df7-n4fhz"] Oct 04 03:25:58 crc kubenswrapper[4742]: W1004 03:25:58.533227 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf99f948a_5fbd_4292_abcd_c8c64992634e.slice/crio-40170802389cf13df2f0d08ff6ab3f7d9d95f650f519d030b3540813bfb74ab8 WatchSource:0}: Error finding container 40170802389cf13df2f0d08ff6ab3f7d9d95f650f519d030b3540813bfb74ab8: Status 404 returned error can't find the container with id 40170802389cf13df2f0d08ff6ab3f7d9d95f650f519d030b3540813bfb74ab8 Oct 04 03:25:58 crc kubenswrapper[4742]: I1004 03:25:58.929200 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-5b74dd9df7-n4fhz" event={"ID":"f99f948a-5fbd-4292-abcd-c8c64992634e","Type":"ContainerStarted","Data":"d27832505d58c92e85678badd92df7dbd4b242a9d8b6b0ee8f155e2276870711"} Oct 04 03:25:58 crc kubenswrapper[4742]: I1004 03:25:58.929676 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-5b74dd9df7-n4fhz" event={"ID":"f99f948a-5fbd-4292-abcd-c8c64992634e","Type":"ContainerStarted","Data":"40170802389cf13df2f0d08ff6ab3f7d9d95f650f519d030b3540813bfb74ab8"} Oct 04 03:25:58 crc kubenswrapper[4742]: I1004 03:25:58.929696 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/keystone-5b74dd9df7-n4fhz" Oct 04 03:25:58 crc kubenswrapper[4742]: I1004 03:25:58.930872 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f47cfc687-xl4hq" event={"ID":"744a78c0-6d93-4a39-bb0c-68f80ff9716b","Type":"ContainerStarted","Data":"ae8eb9d28c826fe8368fd54475b3b4bc777f5ebecd6420f4ae3ebc66b7ab5782"} Oct 04 03:25:58 crc kubenswrapper[4742]: I1004 03:25:58.931032 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f47cfc687-xl4hq" Oct 04 03:25:58 crc kubenswrapper[4742]: I1004 03:25:58.952809 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-5b74dd9df7-n4fhz" podStartSLOduration=1.952785072 podStartE2EDuration="1.952785072s" podCreationTimestamp="2025-10-04 03:25:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:25:58.951097889 +0000 UTC m=+924.883435065" watchObservedRunningTime="2025-10-04 03:25:58.952785072 +0000 UTC m=+924.885122248" Oct 04 03:25:58 crc kubenswrapper[4742]: I1004 03:25:58.984691 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f47cfc687-xl4hq" podStartSLOduration=2.956979494 podStartE2EDuration="4.984672943s" podCreationTimestamp="2025-10-04 03:25:54 +0000 UTC" firstStartedPulling="2025-10-04 03:25:55.6800349 +0000 UTC m=+921.612372086" lastFinishedPulling="2025-10-04 03:25:57.707728359 +0000 UTC m=+923.640065535" observedRunningTime="2025-10-04 03:25:58.9791162 +0000 UTC m=+924.911453376" watchObservedRunningTime="2025-10-04 03:25:58.984672943 +0000 UTC m=+924.917010119" Oct 04 03:26:04 crc kubenswrapper[4742]: I1004 03:26:04.449640 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5994b8459f-9phsb"] Oct 04 03:26:04 crc kubenswrapper[4742]: I1004 03:26:04.451745 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5994b8459f-9phsb" Oct 04 03:26:04 crc kubenswrapper[4742]: I1004 03:26:04.454480 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-gn4hj" Oct 04 03:26:04 crc kubenswrapper[4742]: I1004 03:26:04.457235 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-service-cert" Oct 04 03:26:04 crc kubenswrapper[4742]: I1004 03:26:04.469390 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5994b8459f-9phsb"] Oct 04 03:26:04 crc kubenswrapper[4742]: I1004 03:26:04.494387 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ab3ea796-e1d2-4d21-980c-648daedc36cd-apiservice-cert\") pod \"horizon-operator-controller-manager-5994b8459f-9phsb\" (UID: \"ab3ea796-e1d2-4d21-980c-648daedc36cd\") " pod="openstack-operators/horizon-operator-controller-manager-5994b8459f-9phsb" Oct 04 03:26:04 crc kubenswrapper[4742]: I1004 03:26:04.494454 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrbrt\" (UniqueName: \"kubernetes.io/projected/ab3ea796-e1d2-4d21-980c-648daedc36cd-kube-api-access-lrbrt\") pod \"horizon-operator-controller-manager-5994b8459f-9phsb\" (UID: \"ab3ea796-e1d2-4d21-980c-648daedc36cd\") " pod="openstack-operators/horizon-operator-controller-manager-5994b8459f-9phsb" Oct 04 03:26:04 crc kubenswrapper[4742]: I1004 03:26:04.494532 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ab3ea796-e1d2-4d21-980c-648daedc36cd-webhook-cert\") pod \"horizon-operator-controller-manager-5994b8459f-9phsb\" (UID: \"ab3ea796-e1d2-4d21-980c-648daedc36cd\") " pod="openstack-operators/horizon-operator-controller-manager-5994b8459f-9phsb" Oct 04 03:26:04 crc kubenswrapper[4742]: I1004 03:26:04.595737 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ab3ea796-e1d2-4d21-980c-648daedc36cd-apiservice-cert\") pod \"horizon-operator-controller-manager-5994b8459f-9phsb\" (UID: \"ab3ea796-e1d2-4d21-980c-648daedc36cd\") " pod="openstack-operators/horizon-operator-controller-manager-5994b8459f-9phsb" Oct 04 03:26:04 crc kubenswrapper[4742]: I1004 03:26:04.595785 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrbrt\" (UniqueName: \"kubernetes.io/projected/ab3ea796-e1d2-4d21-980c-648daedc36cd-kube-api-access-lrbrt\") pod \"horizon-operator-controller-manager-5994b8459f-9phsb\" (UID: \"ab3ea796-e1d2-4d21-980c-648daedc36cd\") " pod="openstack-operators/horizon-operator-controller-manager-5994b8459f-9phsb" Oct 04 03:26:04 crc kubenswrapper[4742]: I1004 03:26:04.595842 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ab3ea796-e1d2-4d21-980c-648daedc36cd-webhook-cert\") pod \"horizon-operator-controller-manager-5994b8459f-9phsb\" (UID: \"ab3ea796-e1d2-4d21-980c-648daedc36cd\") " pod="openstack-operators/horizon-operator-controller-manager-5994b8459f-9phsb" Oct 04 03:26:04 crc kubenswrapper[4742]: I1004 03:26:04.603615 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ab3ea796-e1d2-4d21-980c-648daedc36cd-apiservice-cert\") pod \"horizon-operator-controller-manager-5994b8459f-9phsb\" (UID: \"ab3ea796-e1d2-4d21-980c-648daedc36cd\") " pod="openstack-operators/horizon-operator-controller-manager-5994b8459f-9phsb" Oct 04 03:26:04 crc kubenswrapper[4742]: I1004 03:26:04.603878 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ab3ea796-e1d2-4d21-980c-648daedc36cd-webhook-cert\") pod \"horizon-operator-controller-manager-5994b8459f-9phsb\" (UID: \"ab3ea796-e1d2-4d21-980c-648daedc36cd\") " pod="openstack-operators/horizon-operator-controller-manager-5994b8459f-9phsb" Oct 04 03:26:04 crc kubenswrapper[4742]: I1004 03:26:04.615183 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrbrt\" (UniqueName: \"kubernetes.io/projected/ab3ea796-e1d2-4d21-980c-648daedc36cd-kube-api-access-lrbrt\") pod \"horizon-operator-controller-manager-5994b8459f-9phsb\" (UID: \"ab3ea796-e1d2-4d21-980c-648daedc36cd\") " pod="openstack-operators/horizon-operator-controller-manager-5994b8459f-9phsb" Oct 04 03:26:04 crc kubenswrapper[4742]: I1004 03:26:04.770926 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5994b8459f-9phsb" Oct 04 03:26:05 crc kubenswrapper[4742]: I1004 03:26:05.016621 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5994b8459f-9phsb"] Oct 04 03:26:05 crc kubenswrapper[4742]: I1004 03:26:05.133217 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f47cfc687-xl4hq" Oct 04 03:26:05 crc kubenswrapper[4742]: I1004 03:26:05.994308 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5994b8459f-9phsb" event={"ID":"ab3ea796-e1d2-4d21-980c-648daedc36cd","Type":"ContainerStarted","Data":"17179b2e78081293fba3af89892a44dda2cfebd2e7960262daa6626f484eb688"} Oct 04 03:26:07 crc kubenswrapper[4742]: I1004 03:26:07.594121 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Oct 04 03:26:07 crc kubenswrapper[4742]: I1004 03:26:07.597941 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Oct 04 03:26:07 crc kubenswrapper[4742]: I1004 03:26:07.600879 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-files" Oct 04 03:26:07 crc kubenswrapper[4742]: I1004 03:26:07.601093 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-conf" Oct 04 03:26:07 crc kubenswrapper[4742]: I1004 03:26:07.601235 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-swift-dockercfg-xj6df" Oct 04 03:26:07 crc kubenswrapper[4742]: I1004 03:26:07.601369 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-storage-config-data" Oct 04 03:26:07 crc kubenswrapper[4742]: I1004 03:26:07.636060 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Oct 04 03:26:07 crc kubenswrapper[4742]: I1004 03:26:07.662024 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"7e8086e1-26ed-40f6-a3f7-861ec4d232f9\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 03:26:07 crc kubenswrapper[4742]: I1004 03:26:07.662091 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vg2hm\" (UniqueName: \"kubernetes.io/projected/7e8086e1-26ed-40f6-a3f7-861ec4d232f9-kube-api-access-vg2hm\") pod \"swift-storage-0\" (UID: \"7e8086e1-26ed-40f6-a3f7-861ec4d232f9\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 03:26:07 crc kubenswrapper[4742]: I1004 03:26:07.662135 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/7e8086e1-26ed-40f6-a3f7-861ec4d232f9-lock\") pod \"swift-storage-0\" (UID: \"7e8086e1-26ed-40f6-a3f7-861ec4d232f9\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 03:26:07 crc kubenswrapper[4742]: I1004 03:26:07.662162 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/7e8086e1-26ed-40f6-a3f7-861ec4d232f9-cache\") pod \"swift-storage-0\" (UID: \"7e8086e1-26ed-40f6-a3f7-861ec4d232f9\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 03:26:07 crc kubenswrapper[4742]: I1004 03:26:07.662186 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7e8086e1-26ed-40f6-a3f7-861ec4d232f9-etc-swift\") pod \"swift-storage-0\" (UID: \"7e8086e1-26ed-40f6-a3f7-861ec4d232f9\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 03:26:07 crc kubenswrapper[4742]: I1004 03:26:07.764035 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"7e8086e1-26ed-40f6-a3f7-861ec4d232f9\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 03:26:07 crc kubenswrapper[4742]: I1004 03:26:07.764119 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vg2hm\" (UniqueName: \"kubernetes.io/projected/7e8086e1-26ed-40f6-a3f7-861ec4d232f9-kube-api-access-vg2hm\") pod \"swift-storage-0\" (UID: \"7e8086e1-26ed-40f6-a3f7-861ec4d232f9\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 03:26:07 crc kubenswrapper[4742]: I1004 03:26:07.764165 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/7e8086e1-26ed-40f6-a3f7-861ec4d232f9-lock\") pod \"swift-storage-0\" (UID: \"7e8086e1-26ed-40f6-a3f7-861ec4d232f9\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 03:26:07 crc kubenswrapper[4742]: I1004 03:26:07.764210 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/7e8086e1-26ed-40f6-a3f7-861ec4d232f9-cache\") pod \"swift-storage-0\" (UID: \"7e8086e1-26ed-40f6-a3f7-861ec4d232f9\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 03:26:07 crc kubenswrapper[4742]: I1004 03:26:07.764243 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7e8086e1-26ed-40f6-a3f7-861ec4d232f9-etc-swift\") pod \"swift-storage-0\" (UID: \"7e8086e1-26ed-40f6-a3f7-861ec4d232f9\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 03:26:07 crc kubenswrapper[4742]: E1004 03:26:07.764410 4742 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Oct 04 03:26:07 crc kubenswrapper[4742]: E1004 03:26:07.764427 4742 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Oct 04 03:26:07 crc kubenswrapper[4742]: E1004 03:26:07.764483 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7e8086e1-26ed-40f6-a3f7-861ec4d232f9-etc-swift podName:7e8086e1-26ed-40f6-a3f7-861ec4d232f9 nodeName:}" failed. No retries permitted until 2025-10-04 03:26:08.264464312 +0000 UTC m=+934.196801488 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7e8086e1-26ed-40f6-a3f7-861ec4d232f9-etc-swift") pod "swift-storage-0" (UID: "7e8086e1-26ed-40f6-a3f7-861ec4d232f9") : configmap "swift-ring-files" not found Oct 04 03:26:07 crc kubenswrapper[4742]: I1004 03:26:07.764522 4742 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"7e8086e1-26ed-40f6-a3f7-861ec4d232f9\") device mount path \"/mnt/openstack/pv06\"" pod="glance-kuttl-tests/swift-storage-0" Oct 04 03:26:07 crc kubenswrapper[4742]: I1004 03:26:07.764928 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/7e8086e1-26ed-40f6-a3f7-861ec4d232f9-lock\") pod \"swift-storage-0\" (UID: \"7e8086e1-26ed-40f6-a3f7-861ec4d232f9\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 03:26:07 crc kubenswrapper[4742]: I1004 03:26:07.764956 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/7e8086e1-26ed-40f6-a3f7-861ec4d232f9-cache\") pod \"swift-storage-0\" (UID: \"7e8086e1-26ed-40f6-a3f7-861ec4d232f9\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 03:26:07 crc kubenswrapper[4742]: I1004 03:26:07.786756 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"7e8086e1-26ed-40f6-a3f7-861ec4d232f9\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 03:26:07 crc kubenswrapper[4742]: I1004 03:26:07.797653 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vg2hm\" (UniqueName: \"kubernetes.io/projected/7e8086e1-26ed-40f6-a3f7-861ec4d232f9-kube-api-access-vg2hm\") pod \"swift-storage-0\" (UID: \"7e8086e1-26ed-40f6-a3f7-861ec4d232f9\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 03:26:08 crc kubenswrapper[4742]: I1004 03:26:08.009980 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5994b8459f-9phsb" event={"ID":"ab3ea796-e1d2-4d21-980c-648daedc36cd","Type":"ContainerStarted","Data":"ddd321357eae97954f5f530890c041d5c662805ed787626fa83e992164779f84"} Oct 04 03:26:08 crc kubenswrapper[4742]: I1004 03:26:08.010040 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5994b8459f-9phsb" event={"ID":"ab3ea796-e1d2-4d21-980c-648daedc36cd","Type":"ContainerStarted","Data":"36c0b00c92b6d858e2dc9a1c038632164bc3799b4b4204fb5d58ae4df8e8328b"} Oct 04 03:26:08 crc kubenswrapper[4742]: I1004 03:26:08.010211 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5994b8459f-9phsb" Oct 04 03:26:08 crc kubenswrapper[4742]: I1004 03:26:08.055060 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5994b8459f-9phsb" podStartSLOduration=2.006099567 podStartE2EDuration="4.055042841s" podCreationTimestamp="2025-10-04 03:26:04 +0000 UTC" firstStartedPulling="2025-10-04 03:26:05.030141858 +0000 UTC m=+930.962479034" lastFinishedPulling="2025-10-04 03:26:07.079085132 +0000 UTC m=+933.011422308" observedRunningTime="2025-10-04 03:26:08.0538477 +0000 UTC m=+933.986184866" watchObservedRunningTime="2025-10-04 03:26:08.055042841 +0000 UTC m=+933.987380007" Oct 04 03:26:08 crc kubenswrapper[4742]: I1004 03:26:08.271517 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7e8086e1-26ed-40f6-a3f7-861ec4d232f9-etc-swift\") pod \"swift-storage-0\" (UID: \"7e8086e1-26ed-40f6-a3f7-861ec4d232f9\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 03:26:08 crc kubenswrapper[4742]: E1004 03:26:08.271879 4742 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Oct 04 03:26:08 crc kubenswrapper[4742]: E1004 03:26:08.271923 4742 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Oct 04 03:26:08 crc kubenswrapper[4742]: E1004 03:26:08.272036 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7e8086e1-26ed-40f6-a3f7-861ec4d232f9-etc-swift podName:7e8086e1-26ed-40f6-a3f7-861ec4d232f9 nodeName:}" failed. No retries permitted until 2025-10-04 03:26:09.271979625 +0000 UTC m=+935.204316801 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7e8086e1-26ed-40f6-a3f7-861ec4d232f9-etc-swift") pod "swift-storage-0" (UID: "7e8086e1-26ed-40f6-a3f7-861ec4d232f9") : configmap "swift-ring-files" not found Oct 04 03:26:09 crc kubenswrapper[4742]: I1004 03:26:09.197574 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-index-mw675"] Oct 04 03:26:09 crc kubenswrapper[4742]: I1004 03:26:09.198427 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-mw675" Oct 04 03:26:09 crc kubenswrapper[4742]: I1004 03:26:09.201491 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-index-dockercfg-cs8hw" Oct 04 03:26:09 crc kubenswrapper[4742]: I1004 03:26:09.216213 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-mw675"] Oct 04 03:26:09 crc kubenswrapper[4742]: I1004 03:26:09.291627 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7e8086e1-26ed-40f6-a3f7-861ec4d232f9-etc-swift\") pod \"swift-storage-0\" (UID: \"7e8086e1-26ed-40f6-a3f7-861ec4d232f9\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 03:26:09 crc kubenswrapper[4742]: I1004 03:26:09.291808 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7mdc\" (UniqueName: \"kubernetes.io/projected/985d05d8-7796-45a6-b897-cce338183921-kube-api-access-d7mdc\") pod \"glance-operator-index-mw675\" (UID: \"985d05d8-7796-45a6-b897-cce338183921\") " pod="openstack-operators/glance-operator-index-mw675" Oct 04 03:26:09 crc kubenswrapper[4742]: E1004 03:26:09.291841 4742 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Oct 04 03:26:09 crc kubenswrapper[4742]: E1004 03:26:09.291869 4742 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Oct 04 03:26:09 crc kubenswrapper[4742]: E1004 03:26:09.291923 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7e8086e1-26ed-40f6-a3f7-861ec4d232f9-etc-swift podName:7e8086e1-26ed-40f6-a3f7-861ec4d232f9 nodeName:}" failed. No retries permitted until 2025-10-04 03:26:11.291901705 +0000 UTC m=+937.224238881 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7e8086e1-26ed-40f6-a3f7-861ec4d232f9-etc-swift") pod "swift-storage-0" (UID: "7e8086e1-26ed-40f6-a3f7-861ec4d232f9") : configmap "swift-ring-files" not found Oct 04 03:26:09 crc kubenswrapper[4742]: I1004 03:26:09.393795 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7mdc\" (UniqueName: \"kubernetes.io/projected/985d05d8-7796-45a6-b897-cce338183921-kube-api-access-d7mdc\") pod \"glance-operator-index-mw675\" (UID: \"985d05d8-7796-45a6-b897-cce338183921\") " pod="openstack-operators/glance-operator-index-mw675" Oct 04 03:26:09 crc kubenswrapper[4742]: I1004 03:26:09.432402 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7mdc\" (UniqueName: \"kubernetes.io/projected/985d05d8-7796-45a6-b897-cce338183921-kube-api-access-d7mdc\") pod \"glance-operator-index-mw675\" (UID: \"985d05d8-7796-45a6-b897-cce338183921\") " pod="openstack-operators/glance-operator-index-mw675" Oct 04 03:26:09 crc kubenswrapper[4742]: I1004 03:26:09.532965 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-mw675" Oct 04 03:26:10 crc kubenswrapper[4742]: I1004 03:26:10.032386 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-mw675"] Oct 04 03:26:10 crc kubenswrapper[4742]: W1004 03:26:10.049749 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod985d05d8_7796_45a6_b897_cce338183921.slice/crio-43f3557940d1643c6ea1f486ac3f357a0037be9617cebd025800c86fd474631e WatchSource:0}: Error finding container 43f3557940d1643c6ea1f486ac3f357a0037be9617cebd025800c86fd474631e: Status 404 returned error can't find the container with id 43f3557940d1643c6ea1f486ac3f357a0037be9617cebd025800c86fd474631e Oct 04 03:26:11 crc kubenswrapper[4742]: I1004 03:26:11.039102 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-mw675" event={"ID":"985d05d8-7796-45a6-b897-cce338183921","Type":"ContainerStarted","Data":"43f3557940d1643c6ea1f486ac3f357a0037be9617cebd025800c86fd474631e"} Oct 04 03:26:11 crc kubenswrapper[4742]: I1004 03:26:11.327575 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7e8086e1-26ed-40f6-a3f7-861ec4d232f9-etc-swift\") pod \"swift-storage-0\" (UID: \"7e8086e1-26ed-40f6-a3f7-861ec4d232f9\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 03:26:11 crc kubenswrapper[4742]: E1004 03:26:11.327890 4742 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Oct 04 03:26:11 crc kubenswrapper[4742]: E1004 03:26:11.327936 4742 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Oct 04 03:26:11 crc kubenswrapper[4742]: E1004 03:26:11.328029 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7e8086e1-26ed-40f6-a3f7-861ec4d232f9-etc-swift podName:7e8086e1-26ed-40f6-a3f7-861ec4d232f9 nodeName:}" failed. No retries permitted until 2025-10-04 03:26:15.328000798 +0000 UTC m=+941.260338014 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7e8086e1-26ed-40f6-a3f7-861ec4d232f9-etc-swift") pod "swift-storage-0" (UID: "7e8086e1-26ed-40f6-a3f7-861ec4d232f9") : configmap "swift-ring-files" not found Oct 04 03:26:12 crc kubenswrapper[4742]: I1004 03:26:12.050579 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-mw675" event={"ID":"985d05d8-7796-45a6-b897-cce338183921","Type":"ContainerStarted","Data":"d80c5d2467ddecfc35cf4957464ac7a50d23abf807b9cfe449e5e3ff0ec5e034"} Oct 04 03:26:12 crc kubenswrapper[4742]: I1004 03:26:12.080255 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-index-mw675" podStartSLOduration=1.366988445 podStartE2EDuration="3.080233309s" podCreationTimestamp="2025-10-04 03:26:09 +0000 UTC" firstStartedPulling="2025-10-04 03:26:10.054968104 +0000 UTC m=+935.987305310" lastFinishedPulling="2025-10-04 03:26:11.768212958 +0000 UTC m=+937.700550174" observedRunningTime="2025-10-04 03:26:12.066479745 +0000 UTC m=+937.998816941" watchObservedRunningTime="2025-10-04 03:26:12.080233309 +0000 UTC m=+938.012570495" Oct 04 03:26:13 crc kubenswrapper[4742]: I1004 03:26:13.395237 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/glance-operator-index-mw675"] Oct 04 03:26:14 crc kubenswrapper[4742]: I1004 03:26:14.008154 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-index-dfskt"] Oct 04 03:26:14 crc kubenswrapper[4742]: I1004 03:26:14.010431 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-dfskt" Oct 04 03:26:14 crc kubenswrapper[4742]: I1004 03:26:14.030000 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-dfskt"] Oct 04 03:26:14 crc kubenswrapper[4742]: I1004 03:26:14.076125 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/glance-operator-index-mw675" podUID="985d05d8-7796-45a6-b897-cce338183921" containerName="registry-server" containerID="cri-o://d80c5d2467ddecfc35cf4957464ac7a50d23abf807b9cfe449e5e3ff0ec5e034" gracePeriod=2 Oct 04 03:26:14 crc kubenswrapper[4742]: I1004 03:26:14.078237 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nctwh\" (UniqueName: \"kubernetes.io/projected/9f5106a7-ade4-48c9-9b96-95db975f8a99-kube-api-access-nctwh\") pod \"glance-operator-index-dfskt\" (UID: \"9f5106a7-ade4-48c9-9b96-95db975f8a99\") " pod="openstack-operators/glance-operator-index-dfskt" Oct 04 03:26:14 crc kubenswrapper[4742]: I1004 03:26:14.179752 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nctwh\" (UniqueName: \"kubernetes.io/projected/9f5106a7-ade4-48c9-9b96-95db975f8a99-kube-api-access-nctwh\") pod \"glance-operator-index-dfskt\" (UID: \"9f5106a7-ade4-48c9-9b96-95db975f8a99\") " pod="openstack-operators/glance-operator-index-dfskt" Oct 04 03:26:14 crc kubenswrapper[4742]: I1004 03:26:14.215138 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nctwh\" (UniqueName: \"kubernetes.io/projected/9f5106a7-ade4-48c9-9b96-95db975f8a99-kube-api-access-nctwh\") pod \"glance-operator-index-dfskt\" (UID: \"9f5106a7-ade4-48c9-9b96-95db975f8a99\") " pod="openstack-operators/glance-operator-index-dfskt" Oct 04 03:26:14 crc kubenswrapper[4742]: I1004 03:26:14.357756 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-dfskt" Oct 04 03:26:14 crc kubenswrapper[4742]: I1004 03:26:14.637107 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-dfskt"] Oct 04 03:26:14 crc kubenswrapper[4742]: I1004 03:26:14.642005 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-mw675" Oct 04 03:26:14 crc kubenswrapper[4742]: W1004 03:26:14.646325 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f5106a7_ade4_48c9_9b96_95db975f8a99.slice/crio-a2571caa2991429ef0ed1aea51b93a6877e63cfaaa356bf3a47a6c11a99bbfae WatchSource:0}: Error finding container a2571caa2991429ef0ed1aea51b93a6877e63cfaaa356bf3a47a6c11a99bbfae: Status 404 returned error can't find the container with id a2571caa2991429ef0ed1aea51b93a6877e63cfaaa356bf3a47a6c11a99bbfae Oct 04 03:26:14 crc kubenswrapper[4742]: I1004 03:26:14.687109 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d7mdc\" (UniqueName: \"kubernetes.io/projected/985d05d8-7796-45a6-b897-cce338183921-kube-api-access-d7mdc\") pod \"985d05d8-7796-45a6-b897-cce338183921\" (UID: \"985d05d8-7796-45a6-b897-cce338183921\") " Oct 04 03:26:14 crc kubenswrapper[4742]: I1004 03:26:14.694614 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/985d05d8-7796-45a6-b897-cce338183921-kube-api-access-d7mdc" (OuterVolumeSpecName: "kube-api-access-d7mdc") pod "985d05d8-7796-45a6-b897-cce338183921" (UID: "985d05d8-7796-45a6-b897-cce338183921"). InnerVolumeSpecName "kube-api-access-d7mdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:26:14 crc kubenswrapper[4742]: I1004 03:26:14.780678 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5994b8459f-9phsb" Oct 04 03:26:14 crc kubenswrapper[4742]: I1004 03:26:14.788378 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d7mdc\" (UniqueName: \"kubernetes.io/projected/985d05d8-7796-45a6-b897-cce338183921-kube-api-access-d7mdc\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:15 crc kubenswrapper[4742]: I1004 03:26:15.111001 4742 generic.go:334] "Generic (PLEG): container finished" podID="985d05d8-7796-45a6-b897-cce338183921" containerID="d80c5d2467ddecfc35cf4957464ac7a50d23abf807b9cfe449e5e3ff0ec5e034" exitCode=0 Oct 04 03:26:15 crc kubenswrapper[4742]: I1004 03:26:15.111062 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-mw675" Oct 04 03:26:15 crc kubenswrapper[4742]: I1004 03:26:15.111079 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-mw675" event={"ID":"985d05d8-7796-45a6-b897-cce338183921","Type":"ContainerDied","Data":"d80c5d2467ddecfc35cf4957464ac7a50d23abf807b9cfe449e5e3ff0ec5e034"} Oct 04 03:26:15 crc kubenswrapper[4742]: I1004 03:26:15.111140 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-mw675" event={"ID":"985d05d8-7796-45a6-b897-cce338183921","Type":"ContainerDied","Data":"43f3557940d1643c6ea1f486ac3f357a0037be9617cebd025800c86fd474631e"} Oct 04 03:26:15 crc kubenswrapper[4742]: I1004 03:26:15.111165 4742 scope.go:117] "RemoveContainer" containerID="d80c5d2467ddecfc35cf4957464ac7a50d23abf807b9cfe449e5e3ff0ec5e034" Oct 04 03:26:15 crc kubenswrapper[4742]: I1004 03:26:15.114346 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-dfskt" event={"ID":"9f5106a7-ade4-48c9-9b96-95db975f8a99","Type":"ContainerStarted","Data":"4bd31d420240077eb7c6528d801a3159481e36719152305813b19f07d0f8efc3"} Oct 04 03:26:15 crc kubenswrapper[4742]: I1004 03:26:15.114388 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-dfskt" event={"ID":"9f5106a7-ade4-48c9-9b96-95db975f8a99","Type":"ContainerStarted","Data":"a2571caa2991429ef0ed1aea51b93a6877e63cfaaa356bf3a47a6c11a99bbfae"} Oct 04 03:26:15 crc kubenswrapper[4742]: I1004 03:26:15.136846 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-index-dfskt" podStartSLOduration=2.082979891 podStartE2EDuration="2.136818607s" podCreationTimestamp="2025-10-04 03:26:13 +0000 UTC" firstStartedPulling="2025-10-04 03:26:14.652579134 +0000 UTC m=+940.584916310" lastFinishedPulling="2025-10-04 03:26:14.70641784 +0000 UTC m=+940.638755026" observedRunningTime="2025-10-04 03:26:15.132872566 +0000 UTC m=+941.065209752" watchObservedRunningTime="2025-10-04 03:26:15.136818607 +0000 UTC m=+941.069155823" Oct 04 03:26:15 crc kubenswrapper[4742]: I1004 03:26:15.138187 4742 scope.go:117] "RemoveContainer" containerID="d80c5d2467ddecfc35cf4957464ac7a50d23abf807b9cfe449e5e3ff0ec5e034" Oct 04 03:26:15 crc kubenswrapper[4742]: E1004 03:26:15.139822 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d80c5d2467ddecfc35cf4957464ac7a50d23abf807b9cfe449e5e3ff0ec5e034\": container with ID starting with d80c5d2467ddecfc35cf4957464ac7a50d23abf807b9cfe449e5e3ff0ec5e034 not found: ID does not exist" containerID="d80c5d2467ddecfc35cf4957464ac7a50d23abf807b9cfe449e5e3ff0ec5e034" Oct 04 03:26:15 crc kubenswrapper[4742]: I1004 03:26:15.139891 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d80c5d2467ddecfc35cf4957464ac7a50d23abf807b9cfe449e5e3ff0ec5e034"} err="failed to get container status \"d80c5d2467ddecfc35cf4957464ac7a50d23abf807b9cfe449e5e3ff0ec5e034\": rpc error: code = NotFound desc = could not find container \"d80c5d2467ddecfc35cf4957464ac7a50d23abf807b9cfe449e5e3ff0ec5e034\": container with ID starting with d80c5d2467ddecfc35cf4957464ac7a50d23abf807b9cfe449e5e3ff0ec5e034 not found: ID does not exist" Oct 04 03:26:15 crc kubenswrapper[4742]: I1004 03:26:15.155363 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/glance-operator-index-mw675"] Oct 04 03:26:15 crc kubenswrapper[4742]: I1004 03:26:15.164740 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/glance-operator-index-mw675"] Oct 04 03:26:15 crc kubenswrapper[4742]: I1004 03:26:15.405445 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7e8086e1-26ed-40f6-a3f7-861ec4d232f9-etc-swift\") pod \"swift-storage-0\" (UID: \"7e8086e1-26ed-40f6-a3f7-861ec4d232f9\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 03:26:15 crc kubenswrapper[4742]: E1004 03:26:15.405671 4742 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Oct 04 03:26:15 crc kubenswrapper[4742]: E1004 03:26:15.405704 4742 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Oct 04 03:26:15 crc kubenswrapper[4742]: E1004 03:26:15.405758 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7e8086e1-26ed-40f6-a3f7-861ec4d232f9-etc-swift podName:7e8086e1-26ed-40f6-a3f7-861ec4d232f9 nodeName:}" failed. No retries permitted until 2025-10-04 03:26:23.405739829 +0000 UTC m=+949.338077015 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7e8086e1-26ed-40f6-a3f7-861ec4d232f9-etc-swift") pod "swift-storage-0" (UID: "7e8086e1-26ed-40f6-a3f7-861ec4d232f9") : configmap "swift-ring-files" not found Oct 04 03:26:16 crc kubenswrapper[4742]: I1004 03:26:16.874954 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="985d05d8-7796-45a6-b897-cce338183921" path="/var/lib/kubelet/pods/985d05d8-7796-45a6-b897-cce338183921/volumes" Oct 04 03:26:23 crc kubenswrapper[4742]: I1004 03:26:23.433504 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7e8086e1-26ed-40f6-a3f7-861ec4d232f9-etc-swift\") pod \"swift-storage-0\" (UID: \"7e8086e1-26ed-40f6-a3f7-861ec4d232f9\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 03:26:23 crc kubenswrapper[4742]: E1004 03:26:23.433687 4742 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Oct 04 03:26:23 crc kubenswrapper[4742]: E1004 03:26:23.434589 4742 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Oct 04 03:26:23 crc kubenswrapper[4742]: E1004 03:26:23.434647 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7e8086e1-26ed-40f6-a3f7-861ec4d232f9-etc-swift podName:7e8086e1-26ed-40f6-a3f7-861ec4d232f9 nodeName:}" failed. No retries permitted until 2025-10-04 03:26:39.434628431 +0000 UTC m=+965.366965617 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7e8086e1-26ed-40f6-a3f7-861ec4d232f9-etc-swift") pod "swift-storage-0" (UID: "7e8086e1-26ed-40f6-a3f7-861ec4d232f9") : configmap "swift-ring-files" not found Oct 04 03:26:24 crc kubenswrapper[4742]: I1004 03:26:24.358402 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-index-dfskt" Oct 04 03:26:24 crc kubenswrapper[4742]: I1004 03:26:24.358460 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/glance-operator-index-dfskt" Oct 04 03:26:24 crc kubenswrapper[4742]: I1004 03:26:24.407963 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/glance-operator-index-dfskt" Oct 04 03:26:24 crc kubenswrapper[4742]: I1004 03:26:24.848359 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-jvgv2"] Oct 04 03:26:24 crc kubenswrapper[4742]: E1004 03:26:24.848677 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="985d05d8-7796-45a6-b897-cce338183921" containerName="registry-server" Oct 04 03:26:24 crc kubenswrapper[4742]: I1004 03:26:24.848697 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="985d05d8-7796-45a6-b897-cce338183921" containerName="registry-server" Oct 04 03:26:24 crc kubenswrapper[4742]: I1004 03:26:24.848875 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="985d05d8-7796-45a6-b897-cce338183921" containerName="registry-server" Oct 04 03:26:24 crc kubenswrapper[4742]: I1004 03:26:24.849429 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-jvgv2" Oct 04 03:26:24 crc kubenswrapper[4742]: I1004 03:26:24.852722 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-proxy-config-data" Oct 04 03:26:24 crc kubenswrapper[4742]: I1004 03:26:24.852953 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-scripts" Oct 04 03:26:24 crc kubenswrapper[4742]: I1004 03:26:24.854543 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-config-data" Oct 04 03:26:24 crc kubenswrapper[4742]: I1004 03:26:24.873644 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-jvgv2"] Oct 04 03:26:24 crc kubenswrapper[4742]: I1004 03:26:24.960997 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/70411213-953f-4207-a31a-4dd2214efd9e-etc-swift\") pod \"swift-ring-rebalance-jvgv2\" (UID: \"70411213-953f-4207-a31a-4dd2214efd9e\") " pod="glance-kuttl-tests/swift-ring-rebalance-jvgv2" Oct 04 03:26:24 crc kubenswrapper[4742]: I1004 03:26:24.961555 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/70411213-953f-4207-a31a-4dd2214efd9e-ring-data-devices\") pod \"swift-ring-rebalance-jvgv2\" (UID: \"70411213-953f-4207-a31a-4dd2214efd9e\") " pod="glance-kuttl-tests/swift-ring-rebalance-jvgv2" Oct 04 03:26:24 crc kubenswrapper[4742]: I1004 03:26:24.961693 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/70411213-953f-4207-a31a-4dd2214efd9e-dispersionconf\") pod \"swift-ring-rebalance-jvgv2\" (UID: \"70411213-953f-4207-a31a-4dd2214efd9e\") " pod="glance-kuttl-tests/swift-ring-rebalance-jvgv2" Oct 04 03:26:24 crc kubenswrapper[4742]: I1004 03:26:24.961953 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/70411213-953f-4207-a31a-4dd2214efd9e-scripts\") pod \"swift-ring-rebalance-jvgv2\" (UID: \"70411213-953f-4207-a31a-4dd2214efd9e\") " pod="glance-kuttl-tests/swift-ring-rebalance-jvgv2" Oct 04 03:26:24 crc kubenswrapper[4742]: I1004 03:26:24.962152 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zhrn\" (UniqueName: \"kubernetes.io/projected/70411213-953f-4207-a31a-4dd2214efd9e-kube-api-access-5zhrn\") pod \"swift-ring-rebalance-jvgv2\" (UID: \"70411213-953f-4207-a31a-4dd2214efd9e\") " pod="glance-kuttl-tests/swift-ring-rebalance-jvgv2" Oct 04 03:26:24 crc kubenswrapper[4742]: I1004 03:26:24.962301 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/70411213-953f-4207-a31a-4dd2214efd9e-swiftconf\") pod \"swift-ring-rebalance-jvgv2\" (UID: \"70411213-953f-4207-a31a-4dd2214efd9e\") " pod="glance-kuttl-tests/swift-ring-rebalance-jvgv2" Oct 04 03:26:25 crc kubenswrapper[4742]: I1004 03:26:25.064482 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/70411213-953f-4207-a31a-4dd2214efd9e-etc-swift\") pod \"swift-ring-rebalance-jvgv2\" (UID: \"70411213-953f-4207-a31a-4dd2214efd9e\") " pod="glance-kuttl-tests/swift-ring-rebalance-jvgv2" Oct 04 03:26:25 crc kubenswrapper[4742]: I1004 03:26:25.064602 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/70411213-953f-4207-a31a-4dd2214efd9e-ring-data-devices\") pod \"swift-ring-rebalance-jvgv2\" (UID: \"70411213-953f-4207-a31a-4dd2214efd9e\") " pod="glance-kuttl-tests/swift-ring-rebalance-jvgv2" Oct 04 03:26:25 crc kubenswrapper[4742]: I1004 03:26:25.064636 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/70411213-953f-4207-a31a-4dd2214efd9e-dispersionconf\") pod \"swift-ring-rebalance-jvgv2\" (UID: \"70411213-953f-4207-a31a-4dd2214efd9e\") " pod="glance-kuttl-tests/swift-ring-rebalance-jvgv2" Oct 04 03:26:25 crc kubenswrapper[4742]: I1004 03:26:25.064661 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/70411213-953f-4207-a31a-4dd2214efd9e-scripts\") pod \"swift-ring-rebalance-jvgv2\" (UID: \"70411213-953f-4207-a31a-4dd2214efd9e\") " pod="glance-kuttl-tests/swift-ring-rebalance-jvgv2" Oct 04 03:26:25 crc kubenswrapper[4742]: I1004 03:26:25.064715 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zhrn\" (UniqueName: \"kubernetes.io/projected/70411213-953f-4207-a31a-4dd2214efd9e-kube-api-access-5zhrn\") pod \"swift-ring-rebalance-jvgv2\" (UID: \"70411213-953f-4207-a31a-4dd2214efd9e\") " pod="glance-kuttl-tests/swift-ring-rebalance-jvgv2" Oct 04 03:26:25 crc kubenswrapper[4742]: I1004 03:26:25.064754 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/70411213-953f-4207-a31a-4dd2214efd9e-swiftconf\") pod \"swift-ring-rebalance-jvgv2\" (UID: \"70411213-953f-4207-a31a-4dd2214efd9e\") " pod="glance-kuttl-tests/swift-ring-rebalance-jvgv2" Oct 04 03:26:25 crc kubenswrapper[4742]: I1004 03:26:25.066150 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/70411213-953f-4207-a31a-4dd2214efd9e-etc-swift\") pod \"swift-ring-rebalance-jvgv2\" (UID: \"70411213-953f-4207-a31a-4dd2214efd9e\") " pod="glance-kuttl-tests/swift-ring-rebalance-jvgv2" Oct 04 03:26:25 crc kubenswrapper[4742]: I1004 03:26:25.066415 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/70411213-953f-4207-a31a-4dd2214efd9e-scripts\") pod \"swift-ring-rebalance-jvgv2\" (UID: \"70411213-953f-4207-a31a-4dd2214efd9e\") " pod="glance-kuttl-tests/swift-ring-rebalance-jvgv2" Oct 04 03:26:25 crc kubenswrapper[4742]: I1004 03:26:25.066459 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/70411213-953f-4207-a31a-4dd2214efd9e-ring-data-devices\") pod \"swift-ring-rebalance-jvgv2\" (UID: \"70411213-953f-4207-a31a-4dd2214efd9e\") " pod="glance-kuttl-tests/swift-ring-rebalance-jvgv2" Oct 04 03:26:25 crc kubenswrapper[4742]: I1004 03:26:25.074740 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/70411213-953f-4207-a31a-4dd2214efd9e-dispersionconf\") pod \"swift-ring-rebalance-jvgv2\" (UID: \"70411213-953f-4207-a31a-4dd2214efd9e\") " pod="glance-kuttl-tests/swift-ring-rebalance-jvgv2" Oct 04 03:26:25 crc kubenswrapper[4742]: I1004 03:26:25.077876 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/70411213-953f-4207-a31a-4dd2214efd9e-swiftconf\") pod \"swift-ring-rebalance-jvgv2\" (UID: \"70411213-953f-4207-a31a-4dd2214efd9e\") " pod="glance-kuttl-tests/swift-ring-rebalance-jvgv2" Oct 04 03:26:25 crc kubenswrapper[4742]: I1004 03:26:25.093121 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zhrn\" (UniqueName: \"kubernetes.io/projected/70411213-953f-4207-a31a-4dd2214efd9e-kube-api-access-5zhrn\") pod \"swift-ring-rebalance-jvgv2\" (UID: \"70411213-953f-4207-a31a-4dd2214efd9e\") " pod="glance-kuttl-tests/swift-ring-rebalance-jvgv2" Oct 04 03:26:25 crc kubenswrapper[4742]: I1004 03:26:25.237958 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-jvgv2" Oct 04 03:26:25 crc kubenswrapper[4742]: I1004 03:26:25.262038 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-index-dfskt" Oct 04 03:26:25 crc kubenswrapper[4742]: I1004 03:26:25.544050 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-jvgv2"] Oct 04 03:26:26 crc kubenswrapper[4742]: I1004 03:26:26.228191 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-jvgv2" event={"ID":"70411213-953f-4207-a31a-4dd2214efd9e","Type":"ContainerStarted","Data":"b9a48705119143c59394757b6cb016c26c353e071baa4822cf639d433f166b7c"} Oct 04 03:26:28 crc kubenswrapper[4742]: I1004 03:26:28.266349 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq"] Oct 04 03:26:28 crc kubenswrapper[4742]: I1004 03:26:28.268628 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq" Oct 04 03:26:28 crc kubenswrapper[4742]: I1004 03:26:28.271343 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-625gw" Oct 04 03:26:28 crc kubenswrapper[4742]: I1004 03:26:28.286856 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq"] Oct 04 03:26:28 crc kubenswrapper[4742]: I1004 03:26:28.319397 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b62bb235-84cf-41e7-806e-a3f1f03ee01b-bundle\") pod \"f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq\" (UID: \"b62bb235-84cf-41e7-806e-a3f1f03ee01b\") " pod="openstack-operators/f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq" Oct 04 03:26:28 crc kubenswrapper[4742]: I1004 03:26:28.319447 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b62bb235-84cf-41e7-806e-a3f1f03ee01b-util\") pod \"f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq\" (UID: \"b62bb235-84cf-41e7-806e-a3f1f03ee01b\") " pod="openstack-operators/f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq" Oct 04 03:26:28 crc kubenswrapper[4742]: I1004 03:26:28.319498 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzg5z\" (UniqueName: \"kubernetes.io/projected/b62bb235-84cf-41e7-806e-a3f1f03ee01b-kube-api-access-tzg5z\") pod \"f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq\" (UID: \"b62bb235-84cf-41e7-806e-a3f1f03ee01b\") " pod="openstack-operators/f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq" Oct 04 03:26:28 crc kubenswrapper[4742]: I1004 03:26:28.421356 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b62bb235-84cf-41e7-806e-a3f1f03ee01b-bundle\") pod \"f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq\" (UID: \"b62bb235-84cf-41e7-806e-a3f1f03ee01b\") " pod="openstack-operators/f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq" Oct 04 03:26:28 crc kubenswrapper[4742]: I1004 03:26:28.421421 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b62bb235-84cf-41e7-806e-a3f1f03ee01b-util\") pod \"f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq\" (UID: \"b62bb235-84cf-41e7-806e-a3f1f03ee01b\") " pod="openstack-operators/f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq" Oct 04 03:26:28 crc kubenswrapper[4742]: I1004 03:26:28.421479 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzg5z\" (UniqueName: \"kubernetes.io/projected/b62bb235-84cf-41e7-806e-a3f1f03ee01b-kube-api-access-tzg5z\") pod \"f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq\" (UID: \"b62bb235-84cf-41e7-806e-a3f1f03ee01b\") " pod="openstack-operators/f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq" Oct 04 03:26:28 crc kubenswrapper[4742]: I1004 03:26:28.422323 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b62bb235-84cf-41e7-806e-a3f1f03ee01b-bundle\") pod \"f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq\" (UID: \"b62bb235-84cf-41e7-806e-a3f1f03ee01b\") " pod="openstack-operators/f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq" Oct 04 03:26:28 crc kubenswrapper[4742]: I1004 03:26:28.422389 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b62bb235-84cf-41e7-806e-a3f1f03ee01b-util\") pod \"f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq\" (UID: \"b62bb235-84cf-41e7-806e-a3f1f03ee01b\") " pod="openstack-operators/f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq" Oct 04 03:26:28 crc kubenswrapper[4742]: I1004 03:26:28.444013 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzg5z\" (UniqueName: \"kubernetes.io/projected/b62bb235-84cf-41e7-806e-a3f1f03ee01b-kube-api-access-tzg5z\") pod \"f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq\" (UID: \"b62bb235-84cf-41e7-806e-a3f1f03ee01b\") " pod="openstack-operators/f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq" Oct 04 03:26:28 crc kubenswrapper[4742]: I1004 03:26:28.596935 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq" Oct 04 03:26:29 crc kubenswrapper[4742]: I1004 03:26:29.472820 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/keystone-5b74dd9df7-n4fhz" Oct 04 03:26:33 crc kubenswrapper[4742]: I1004 03:26:33.232818 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq"] Oct 04 03:26:33 crc kubenswrapper[4742]: W1004 03:26:33.236228 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb62bb235_84cf_41e7_806e_a3f1f03ee01b.slice/crio-042dc1c064fa6fa0200500792b712d660236411eb2e555b118981d8531859eba WatchSource:0}: Error finding container 042dc1c064fa6fa0200500792b712d660236411eb2e555b118981d8531859eba: Status 404 returned error can't find the container with id 042dc1c064fa6fa0200500792b712d660236411eb2e555b118981d8531859eba Oct 04 03:26:33 crc kubenswrapper[4742]: I1004 03:26:33.290513 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq" event={"ID":"b62bb235-84cf-41e7-806e-a3f1f03ee01b","Type":"ContainerStarted","Data":"042dc1c064fa6fa0200500792b712d660236411eb2e555b118981d8531859eba"} Oct 04 03:26:33 crc kubenswrapper[4742]: I1004 03:26:33.292095 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-jvgv2" event={"ID":"70411213-953f-4207-a31a-4dd2214efd9e","Type":"ContainerStarted","Data":"1c3dfb4db2a3dc664fe4e6f8e72c95cbb26dff0d23999ded1e4ff5fe62f58772"} Oct 04 03:26:33 crc kubenswrapper[4742]: I1004 03:26:33.317262 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-ring-rebalance-jvgv2" podStartSLOduration=1.8540499499999998 podStartE2EDuration="9.317234153s" podCreationTimestamp="2025-10-04 03:26:24 +0000 UTC" firstStartedPulling="2025-10-04 03:26:25.55760975 +0000 UTC m=+951.489946936" lastFinishedPulling="2025-10-04 03:26:33.020793963 +0000 UTC m=+958.953131139" observedRunningTime="2025-10-04 03:26:33.311788963 +0000 UTC m=+959.244126159" watchObservedRunningTime="2025-10-04 03:26:33.317234153 +0000 UTC m=+959.249571329" Oct 04 03:26:34 crc kubenswrapper[4742]: I1004 03:26:34.309030 4742 generic.go:334] "Generic (PLEG): container finished" podID="b62bb235-84cf-41e7-806e-a3f1f03ee01b" containerID="b6365237d7f14703958142c20f65ecdc1927b9f8aa2412a5d157882f825a2e12" exitCode=0 Oct 04 03:26:34 crc kubenswrapper[4742]: I1004 03:26:34.309096 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq" event={"ID":"b62bb235-84cf-41e7-806e-a3f1f03ee01b","Type":"ContainerDied","Data":"b6365237d7f14703958142c20f65ecdc1927b9f8aa2412a5d157882f825a2e12"} Oct 04 03:26:36 crc kubenswrapper[4742]: I1004 03:26:36.326614 4742 generic.go:334] "Generic (PLEG): container finished" podID="b62bb235-84cf-41e7-806e-a3f1f03ee01b" containerID="b2e25aeab129d30e6554abf7ce6c02544b8bae004490b606bcc0294dea841625" exitCode=0 Oct 04 03:26:36 crc kubenswrapper[4742]: I1004 03:26:36.326752 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq" event={"ID":"b62bb235-84cf-41e7-806e-a3f1f03ee01b","Type":"ContainerDied","Data":"b2e25aeab129d30e6554abf7ce6c02544b8bae004490b606bcc0294dea841625"} Oct 04 03:26:37 crc kubenswrapper[4742]: I1004 03:26:37.340079 4742 generic.go:334] "Generic (PLEG): container finished" podID="b62bb235-84cf-41e7-806e-a3f1f03ee01b" containerID="9c4ffac4e340f15c723bdf0a35df712123fdbed77ba345e50679dbfbfa29d3db" exitCode=0 Oct 04 03:26:37 crc kubenswrapper[4742]: I1004 03:26:37.340163 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq" event={"ID":"b62bb235-84cf-41e7-806e-a3f1f03ee01b","Type":"ContainerDied","Data":"9c4ffac4e340f15c723bdf0a35df712123fdbed77ba345e50679dbfbfa29d3db"} Oct 04 03:26:38 crc kubenswrapper[4742]: I1004 03:26:38.494526 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-proxy-59cb459c9f-hs28w"] Oct 04 03:26:38 crc kubenswrapper[4742]: I1004 03:26:38.496202 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-59cb459c9f-hs28w" Oct 04 03:26:38 crc kubenswrapper[4742]: I1004 03:26:38.508982 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-proxy-59cb459c9f-hs28w"] Oct 04 03:26:38 crc kubenswrapper[4742]: I1004 03:26:38.513296 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5d409f2c-d078-4589-87c8-7e6fb42e667b-run-httpd\") pod \"swift-proxy-59cb459c9f-hs28w\" (UID: \"5d409f2c-d078-4589-87c8-7e6fb42e667b\") " pod="glance-kuttl-tests/swift-proxy-59cb459c9f-hs28w" Oct 04 03:26:38 crc kubenswrapper[4742]: I1004 03:26:38.513398 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5d409f2c-d078-4589-87c8-7e6fb42e667b-log-httpd\") pod \"swift-proxy-59cb459c9f-hs28w\" (UID: \"5d409f2c-d078-4589-87c8-7e6fb42e667b\") " pod="glance-kuttl-tests/swift-proxy-59cb459c9f-hs28w" Oct 04 03:26:38 crc kubenswrapper[4742]: I1004 03:26:38.513426 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtj2n\" (UniqueName: \"kubernetes.io/projected/5d409f2c-d078-4589-87c8-7e6fb42e667b-kube-api-access-rtj2n\") pod \"swift-proxy-59cb459c9f-hs28w\" (UID: \"5d409f2c-d078-4589-87c8-7e6fb42e667b\") " pod="glance-kuttl-tests/swift-proxy-59cb459c9f-hs28w" Oct 04 03:26:38 crc kubenswrapper[4742]: I1004 03:26:38.513472 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5d409f2c-d078-4589-87c8-7e6fb42e667b-etc-swift\") pod \"swift-proxy-59cb459c9f-hs28w\" (UID: \"5d409f2c-d078-4589-87c8-7e6fb42e667b\") " pod="glance-kuttl-tests/swift-proxy-59cb459c9f-hs28w" Oct 04 03:26:38 crc kubenswrapper[4742]: I1004 03:26:38.513501 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d409f2c-d078-4589-87c8-7e6fb42e667b-config-data\") pod \"swift-proxy-59cb459c9f-hs28w\" (UID: \"5d409f2c-d078-4589-87c8-7e6fb42e667b\") " pod="glance-kuttl-tests/swift-proxy-59cb459c9f-hs28w" Oct 04 03:26:38 crc kubenswrapper[4742]: I1004 03:26:38.615150 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5d409f2c-d078-4589-87c8-7e6fb42e667b-run-httpd\") pod \"swift-proxy-59cb459c9f-hs28w\" (UID: \"5d409f2c-d078-4589-87c8-7e6fb42e667b\") " pod="glance-kuttl-tests/swift-proxy-59cb459c9f-hs28w" Oct 04 03:26:38 crc kubenswrapper[4742]: I1004 03:26:38.615771 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5d409f2c-d078-4589-87c8-7e6fb42e667b-log-httpd\") pod \"swift-proxy-59cb459c9f-hs28w\" (UID: \"5d409f2c-d078-4589-87c8-7e6fb42e667b\") " pod="glance-kuttl-tests/swift-proxy-59cb459c9f-hs28w" Oct 04 03:26:38 crc kubenswrapper[4742]: I1004 03:26:38.615812 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtj2n\" (UniqueName: \"kubernetes.io/projected/5d409f2c-d078-4589-87c8-7e6fb42e667b-kube-api-access-rtj2n\") pod \"swift-proxy-59cb459c9f-hs28w\" (UID: \"5d409f2c-d078-4589-87c8-7e6fb42e667b\") " pod="glance-kuttl-tests/swift-proxy-59cb459c9f-hs28w" Oct 04 03:26:38 crc kubenswrapper[4742]: I1004 03:26:38.615874 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5d409f2c-d078-4589-87c8-7e6fb42e667b-etc-swift\") pod \"swift-proxy-59cb459c9f-hs28w\" (UID: \"5d409f2c-d078-4589-87c8-7e6fb42e667b\") " pod="glance-kuttl-tests/swift-proxy-59cb459c9f-hs28w" Oct 04 03:26:38 crc kubenswrapper[4742]: I1004 03:26:38.615915 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d409f2c-d078-4589-87c8-7e6fb42e667b-config-data\") pod \"swift-proxy-59cb459c9f-hs28w\" (UID: \"5d409f2c-d078-4589-87c8-7e6fb42e667b\") " pod="glance-kuttl-tests/swift-proxy-59cb459c9f-hs28w" Oct 04 03:26:38 crc kubenswrapper[4742]: I1004 03:26:38.616194 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5d409f2c-d078-4589-87c8-7e6fb42e667b-run-httpd\") pod \"swift-proxy-59cb459c9f-hs28w\" (UID: \"5d409f2c-d078-4589-87c8-7e6fb42e667b\") " pod="glance-kuttl-tests/swift-proxy-59cb459c9f-hs28w" Oct 04 03:26:38 crc kubenswrapper[4742]: I1004 03:26:38.616887 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5d409f2c-d078-4589-87c8-7e6fb42e667b-log-httpd\") pod \"swift-proxy-59cb459c9f-hs28w\" (UID: \"5d409f2c-d078-4589-87c8-7e6fb42e667b\") " pod="glance-kuttl-tests/swift-proxy-59cb459c9f-hs28w" Oct 04 03:26:38 crc kubenswrapper[4742]: E1004 03:26:38.616997 4742 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Oct 04 03:26:38 crc kubenswrapper[4742]: E1004 03:26:38.617020 4742 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-59cb459c9f-hs28w: configmap "swift-ring-files" not found Oct 04 03:26:38 crc kubenswrapper[4742]: E1004 03:26:38.617068 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5d409f2c-d078-4589-87c8-7e6fb42e667b-etc-swift podName:5d409f2c-d078-4589-87c8-7e6fb42e667b nodeName:}" failed. No retries permitted until 2025-10-04 03:26:39.117050116 +0000 UTC m=+965.049387292 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/5d409f2c-d078-4589-87c8-7e6fb42e667b-etc-swift") pod "swift-proxy-59cb459c9f-hs28w" (UID: "5d409f2c-d078-4589-87c8-7e6fb42e667b") : configmap "swift-ring-files" not found Oct 04 03:26:38 crc kubenswrapper[4742]: I1004 03:26:38.632079 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d409f2c-d078-4589-87c8-7e6fb42e667b-config-data\") pod \"swift-proxy-59cb459c9f-hs28w\" (UID: \"5d409f2c-d078-4589-87c8-7e6fb42e667b\") " pod="glance-kuttl-tests/swift-proxy-59cb459c9f-hs28w" Oct 04 03:26:38 crc kubenswrapper[4742]: I1004 03:26:38.632720 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtj2n\" (UniqueName: \"kubernetes.io/projected/5d409f2c-d078-4589-87c8-7e6fb42e667b-kube-api-access-rtj2n\") pod \"swift-proxy-59cb459c9f-hs28w\" (UID: \"5d409f2c-d078-4589-87c8-7e6fb42e667b\") " pod="glance-kuttl-tests/swift-proxy-59cb459c9f-hs28w" Oct 04 03:26:38 crc kubenswrapper[4742]: I1004 03:26:38.680602 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq" Oct 04 03:26:38 crc kubenswrapper[4742]: I1004 03:26:38.717335 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzg5z\" (UniqueName: \"kubernetes.io/projected/b62bb235-84cf-41e7-806e-a3f1f03ee01b-kube-api-access-tzg5z\") pod \"b62bb235-84cf-41e7-806e-a3f1f03ee01b\" (UID: \"b62bb235-84cf-41e7-806e-a3f1f03ee01b\") " Oct 04 03:26:38 crc kubenswrapper[4742]: I1004 03:26:38.717433 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b62bb235-84cf-41e7-806e-a3f1f03ee01b-util\") pod \"b62bb235-84cf-41e7-806e-a3f1f03ee01b\" (UID: \"b62bb235-84cf-41e7-806e-a3f1f03ee01b\") " Oct 04 03:26:38 crc kubenswrapper[4742]: I1004 03:26:38.717559 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b62bb235-84cf-41e7-806e-a3f1f03ee01b-bundle\") pod \"b62bb235-84cf-41e7-806e-a3f1f03ee01b\" (UID: \"b62bb235-84cf-41e7-806e-a3f1f03ee01b\") " Oct 04 03:26:38 crc kubenswrapper[4742]: I1004 03:26:38.719701 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b62bb235-84cf-41e7-806e-a3f1f03ee01b-bundle" (OuterVolumeSpecName: "bundle") pod "b62bb235-84cf-41e7-806e-a3f1f03ee01b" (UID: "b62bb235-84cf-41e7-806e-a3f1f03ee01b"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:26:38 crc kubenswrapper[4742]: I1004 03:26:38.724101 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b62bb235-84cf-41e7-806e-a3f1f03ee01b-kube-api-access-tzg5z" (OuterVolumeSpecName: "kube-api-access-tzg5z") pod "b62bb235-84cf-41e7-806e-a3f1f03ee01b" (UID: "b62bb235-84cf-41e7-806e-a3f1f03ee01b"). InnerVolumeSpecName "kube-api-access-tzg5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:26:38 crc kubenswrapper[4742]: I1004 03:26:38.732129 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b62bb235-84cf-41e7-806e-a3f1f03ee01b-util" (OuterVolumeSpecName: "util") pod "b62bb235-84cf-41e7-806e-a3f1f03ee01b" (UID: "b62bb235-84cf-41e7-806e-a3f1f03ee01b"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:26:38 crc kubenswrapper[4742]: I1004 03:26:38.819858 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzg5z\" (UniqueName: \"kubernetes.io/projected/b62bb235-84cf-41e7-806e-a3f1f03ee01b-kube-api-access-tzg5z\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:38 crc kubenswrapper[4742]: I1004 03:26:38.819915 4742 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b62bb235-84cf-41e7-806e-a3f1f03ee01b-util\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:38 crc kubenswrapper[4742]: I1004 03:26:38.819932 4742 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b62bb235-84cf-41e7-806e-a3f1f03ee01b-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:39 crc kubenswrapper[4742]: I1004 03:26:39.123956 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5d409f2c-d078-4589-87c8-7e6fb42e667b-etc-swift\") pod \"swift-proxy-59cb459c9f-hs28w\" (UID: \"5d409f2c-d078-4589-87c8-7e6fb42e667b\") " pod="glance-kuttl-tests/swift-proxy-59cb459c9f-hs28w" Oct 04 03:26:39 crc kubenswrapper[4742]: E1004 03:26:39.124239 4742 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Oct 04 03:26:39 crc kubenswrapper[4742]: E1004 03:26:39.124261 4742 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-59cb459c9f-hs28w: configmap "swift-ring-files" not found Oct 04 03:26:39 crc kubenswrapper[4742]: E1004 03:26:39.124342 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5d409f2c-d078-4589-87c8-7e6fb42e667b-etc-swift podName:5d409f2c-d078-4589-87c8-7e6fb42e667b nodeName:}" failed. No retries permitted until 2025-10-04 03:26:40.124317582 +0000 UTC m=+966.056654758 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/5d409f2c-d078-4589-87c8-7e6fb42e667b-etc-swift") pod "swift-proxy-59cb459c9f-hs28w" (UID: "5d409f2c-d078-4589-87c8-7e6fb42e667b") : configmap "swift-ring-files" not found Oct 04 03:26:39 crc kubenswrapper[4742]: I1004 03:26:39.361671 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq" event={"ID":"b62bb235-84cf-41e7-806e-a3f1f03ee01b","Type":"ContainerDied","Data":"042dc1c064fa6fa0200500792b712d660236411eb2e555b118981d8531859eba"} Oct 04 03:26:39 crc kubenswrapper[4742]: I1004 03:26:39.361723 4742 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="042dc1c064fa6fa0200500792b712d660236411eb2e555b118981d8531859eba" Oct 04 03:26:39 crc kubenswrapper[4742]: I1004 03:26:39.361748 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq" Oct 04 03:26:39 crc kubenswrapper[4742]: I1004 03:26:39.529019 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7e8086e1-26ed-40f6-a3f7-861ec4d232f9-etc-swift\") pod \"swift-storage-0\" (UID: \"7e8086e1-26ed-40f6-a3f7-861ec4d232f9\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 03:26:39 crc kubenswrapper[4742]: E1004 03:26:39.529264 4742 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Oct 04 03:26:39 crc kubenswrapper[4742]: E1004 03:26:39.529313 4742 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Oct 04 03:26:39 crc kubenswrapper[4742]: E1004 03:26:39.529383 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7e8086e1-26ed-40f6-a3f7-861ec4d232f9-etc-swift podName:7e8086e1-26ed-40f6-a3f7-861ec4d232f9 nodeName:}" failed. No retries permitted until 2025-10-04 03:27:11.529362807 +0000 UTC m=+997.461699993 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7e8086e1-26ed-40f6-a3f7-861ec4d232f9-etc-swift") pod "swift-storage-0" (UID: "7e8086e1-26ed-40f6-a3f7-861ec4d232f9") : configmap "swift-ring-files" not found Oct 04 03:26:40 crc kubenswrapper[4742]: I1004 03:26:40.141359 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5d409f2c-d078-4589-87c8-7e6fb42e667b-etc-swift\") pod \"swift-proxy-59cb459c9f-hs28w\" (UID: \"5d409f2c-d078-4589-87c8-7e6fb42e667b\") " pod="glance-kuttl-tests/swift-proxy-59cb459c9f-hs28w" Oct 04 03:26:40 crc kubenswrapper[4742]: I1004 03:26:40.165360 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5d409f2c-d078-4589-87c8-7e6fb42e667b-etc-swift\") pod \"swift-proxy-59cb459c9f-hs28w\" (UID: \"5d409f2c-d078-4589-87c8-7e6fb42e667b\") " pod="glance-kuttl-tests/swift-proxy-59cb459c9f-hs28w" Oct 04 03:26:40 crc kubenswrapper[4742]: I1004 03:26:40.331483 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-59cb459c9f-hs28w" Oct 04 03:26:40 crc kubenswrapper[4742]: I1004 03:26:40.374877 4742 generic.go:334] "Generic (PLEG): container finished" podID="70411213-953f-4207-a31a-4dd2214efd9e" containerID="1c3dfb4db2a3dc664fe4e6f8e72c95cbb26dff0d23999ded1e4ff5fe62f58772" exitCode=0 Oct 04 03:26:40 crc kubenswrapper[4742]: I1004 03:26:40.374948 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-jvgv2" event={"ID":"70411213-953f-4207-a31a-4dd2214efd9e","Type":"ContainerDied","Data":"1c3dfb4db2a3dc664fe4e6f8e72c95cbb26dff0d23999ded1e4ff5fe62f58772"} Oct 04 03:26:40 crc kubenswrapper[4742]: I1004 03:26:40.877189 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-proxy-59cb459c9f-hs28w"] Oct 04 03:26:41 crc kubenswrapper[4742]: I1004 03:26:41.388149 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-59cb459c9f-hs28w" event={"ID":"5d409f2c-d078-4589-87c8-7e6fb42e667b","Type":"ContainerStarted","Data":"fffebd0e0c8a0543e9f518f17715e7102d083537fd18e583afa9206c16d4d5b9"} Oct 04 03:26:41 crc kubenswrapper[4742]: I1004 03:26:41.391302 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-59cb459c9f-hs28w" event={"ID":"5d409f2c-d078-4589-87c8-7e6fb42e667b","Type":"ContainerStarted","Data":"10c940e5e1c53b7877c22f25a3d4cf2b73e951c2d63376916bf160b51e432278"} Oct 04 03:26:41 crc kubenswrapper[4742]: I1004 03:26:41.391530 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/swift-proxy-59cb459c9f-hs28w" Oct 04 03:26:41 crc kubenswrapper[4742]: I1004 03:26:41.391737 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/swift-proxy-59cb459c9f-hs28w" Oct 04 03:26:41 crc kubenswrapper[4742]: I1004 03:26:41.392007 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-59cb459c9f-hs28w" event={"ID":"5d409f2c-d078-4589-87c8-7e6fb42e667b","Type":"ContainerStarted","Data":"4a2237d2b0a783f741fa0551e8ab9e15738770ea97c559b718dd5c7e2cd5b0fc"} Oct 04 03:26:41 crc kubenswrapper[4742]: I1004 03:26:41.413484 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-proxy-59cb459c9f-hs28w" podStartSLOduration=3.413451809 podStartE2EDuration="3.413451809s" podCreationTimestamp="2025-10-04 03:26:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:26:41.412123115 +0000 UTC m=+967.344460381" watchObservedRunningTime="2025-10-04 03:26:41.413451809 +0000 UTC m=+967.345789065" Oct 04 03:26:41 crc kubenswrapper[4742]: I1004 03:26:41.749956 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-jvgv2" Oct 04 03:26:41 crc kubenswrapper[4742]: I1004 03:26:41.790701 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/70411213-953f-4207-a31a-4dd2214efd9e-scripts\") pod \"70411213-953f-4207-a31a-4dd2214efd9e\" (UID: \"70411213-953f-4207-a31a-4dd2214efd9e\") " Oct 04 03:26:41 crc kubenswrapper[4742]: I1004 03:26:41.790821 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/70411213-953f-4207-a31a-4dd2214efd9e-dispersionconf\") pod \"70411213-953f-4207-a31a-4dd2214efd9e\" (UID: \"70411213-953f-4207-a31a-4dd2214efd9e\") " Oct 04 03:26:41 crc kubenswrapper[4742]: I1004 03:26:41.790918 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/70411213-953f-4207-a31a-4dd2214efd9e-swiftconf\") pod \"70411213-953f-4207-a31a-4dd2214efd9e\" (UID: \"70411213-953f-4207-a31a-4dd2214efd9e\") " Oct 04 03:26:41 crc kubenswrapper[4742]: I1004 03:26:41.790970 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/70411213-953f-4207-a31a-4dd2214efd9e-etc-swift\") pod \"70411213-953f-4207-a31a-4dd2214efd9e\" (UID: \"70411213-953f-4207-a31a-4dd2214efd9e\") " Oct 04 03:26:41 crc kubenswrapper[4742]: I1004 03:26:41.791025 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/70411213-953f-4207-a31a-4dd2214efd9e-ring-data-devices\") pod \"70411213-953f-4207-a31a-4dd2214efd9e\" (UID: \"70411213-953f-4207-a31a-4dd2214efd9e\") " Oct 04 03:26:41 crc kubenswrapper[4742]: I1004 03:26:41.791069 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zhrn\" (UniqueName: \"kubernetes.io/projected/70411213-953f-4207-a31a-4dd2214efd9e-kube-api-access-5zhrn\") pod \"70411213-953f-4207-a31a-4dd2214efd9e\" (UID: \"70411213-953f-4207-a31a-4dd2214efd9e\") " Oct 04 03:26:41 crc kubenswrapper[4742]: I1004 03:26:41.792713 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70411213-953f-4207-a31a-4dd2214efd9e-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "70411213-953f-4207-a31a-4dd2214efd9e" (UID: "70411213-953f-4207-a31a-4dd2214efd9e"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:26:41 crc kubenswrapper[4742]: I1004 03:26:41.796811 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70411213-953f-4207-a31a-4dd2214efd9e-kube-api-access-5zhrn" (OuterVolumeSpecName: "kube-api-access-5zhrn") pod "70411213-953f-4207-a31a-4dd2214efd9e" (UID: "70411213-953f-4207-a31a-4dd2214efd9e"). InnerVolumeSpecName "kube-api-access-5zhrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:26:41 crc kubenswrapper[4742]: I1004 03:26:41.797471 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70411213-953f-4207-a31a-4dd2214efd9e-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "70411213-953f-4207-a31a-4dd2214efd9e" (UID: "70411213-953f-4207-a31a-4dd2214efd9e"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:26:41 crc kubenswrapper[4742]: I1004 03:26:41.800372 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70411213-953f-4207-a31a-4dd2214efd9e-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "70411213-953f-4207-a31a-4dd2214efd9e" (UID: "70411213-953f-4207-a31a-4dd2214efd9e"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:26:41 crc kubenswrapper[4742]: I1004 03:26:41.821589 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70411213-953f-4207-a31a-4dd2214efd9e-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "70411213-953f-4207-a31a-4dd2214efd9e" (UID: "70411213-953f-4207-a31a-4dd2214efd9e"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:26:41 crc kubenswrapper[4742]: I1004 03:26:41.828575 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70411213-953f-4207-a31a-4dd2214efd9e-scripts" (OuterVolumeSpecName: "scripts") pod "70411213-953f-4207-a31a-4dd2214efd9e" (UID: "70411213-953f-4207-a31a-4dd2214efd9e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:26:41 crc kubenswrapper[4742]: I1004 03:26:41.892424 4742 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/70411213-953f-4207-a31a-4dd2214efd9e-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:41 crc kubenswrapper[4742]: I1004 03:26:41.892469 4742 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/70411213-953f-4207-a31a-4dd2214efd9e-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:41 crc kubenswrapper[4742]: I1004 03:26:41.892483 4742 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/70411213-953f-4207-a31a-4dd2214efd9e-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:41 crc kubenswrapper[4742]: I1004 03:26:41.892493 4742 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/70411213-953f-4207-a31a-4dd2214efd9e-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:41 crc kubenswrapper[4742]: I1004 03:26:41.892508 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zhrn\" (UniqueName: \"kubernetes.io/projected/70411213-953f-4207-a31a-4dd2214efd9e-kube-api-access-5zhrn\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:41 crc kubenswrapper[4742]: I1004 03:26:41.892519 4742 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/70411213-953f-4207-a31a-4dd2214efd9e-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:26:42 crc kubenswrapper[4742]: I1004 03:26:42.395887 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-jvgv2" Oct 04 03:26:42 crc kubenswrapper[4742]: I1004 03:26:42.404496 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-jvgv2" event={"ID":"70411213-953f-4207-a31a-4dd2214efd9e","Type":"ContainerDied","Data":"b9a48705119143c59394757b6cb016c26c353e071baa4822cf639d433f166b7c"} Oct 04 03:26:42 crc kubenswrapper[4742]: I1004 03:26:42.404575 4742 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9a48705119143c59394757b6cb016c26c353e071baa4822cf639d433f166b7c" Oct 04 03:26:49 crc kubenswrapper[4742]: I1004 03:26:49.255387 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-67c6d4bf5f-ngkxb"] Oct 04 03:26:49 crc kubenswrapper[4742]: E1004 03:26:49.256305 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b62bb235-84cf-41e7-806e-a3f1f03ee01b" containerName="extract" Oct 04 03:26:49 crc kubenswrapper[4742]: I1004 03:26:49.256317 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="b62bb235-84cf-41e7-806e-a3f1f03ee01b" containerName="extract" Oct 04 03:26:49 crc kubenswrapper[4742]: E1004 03:26:49.256329 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b62bb235-84cf-41e7-806e-a3f1f03ee01b" containerName="util" Oct 04 03:26:49 crc kubenswrapper[4742]: I1004 03:26:49.256335 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="b62bb235-84cf-41e7-806e-a3f1f03ee01b" containerName="util" Oct 04 03:26:49 crc kubenswrapper[4742]: E1004 03:26:49.256342 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b62bb235-84cf-41e7-806e-a3f1f03ee01b" containerName="pull" Oct 04 03:26:49 crc kubenswrapper[4742]: I1004 03:26:49.256348 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="b62bb235-84cf-41e7-806e-a3f1f03ee01b" containerName="pull" Oct 04 03:26:49 crc kubenswrapper[4742]: E1004 03:26:49.256380 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70411213-953f-4207-a31a-4dd2214efd9e" containerName="swift-ring-rebalance" Oct 04 03:26:49 crc kubenswrapper[4742]: I1004 03:26:49.256387 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="70411213-953f-4207-a31a-4dd2214efd9e" containerName="swift-ring-rebalance" Oct 04 03:26:49 crc kubenswrapper[4742]: I1004 03:26:49.256525 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="70411213-953f-4207-a31a-4dd2214efd9e" containerName="swift-ring-rebalance" Oct 04 03:26:49 crc kubenswrapper[4742]: I1004 03:26:49.256538 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="b62bb235-84cf-41e7-806e-a3f1f03ee01b" containerName="extract" Oct 04 03:26:49 crc kubenswrapper[4742]: I1004 03:26:49.257236 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-67c6d4bf5f-ngkxb" Oct 04 03:26:49 crc kubenswrapper[4742]: I1004 03:26:49.263078 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-service-cert" Oct 04 03:26:49 crc kubenswrapper[4742]: I1004 03:26:49.263697 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-zsdjx" Oct 04 03:26:49 crc kubenswrapper[4742]: I1004 03:26:49.281238 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-67c6d4bf5f-ngkxb"] Oct 04 03:26:49 crc kubenswrapper[4742]: I1004 03:26:49.331566 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kc527\" (UniqueName: \"kubernetes.io/projected/94b82631-9d81-4926-8677-1d9addae8808-kube-api-access-kc527\") pod \"glance-operator-controller-manager-67c6d4bf5f-ngkxb\" (UID: \"94b82631-9d81-4926-8677-1d9addae8808\") " pod="openstack-operators/glance-operator-controller-manager-67c6d4bf5f-ngkxb" Oct 04 03:26:49 crc kubenswrapper[4742]: I1004 03:26:49.331683 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/94b82631-9d81-4926-8677-1d9addae8808-webhook-cert\") pod \"glance-operator-controller-manager-67c6d4bf5f-ngkxb\" (UID: \"94b82631-9d81-4926-8677-1d9addae8808\") " pod="openstack-operators/glance-operator-controller-manager-67c6d4bf5f-ngkxb" Oct 04 03:26:49 crc kubenswrapper[4742]: I1004 03:26:49.331729 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/94b82631-9d81-4926-8677-1d9addae8808-apiservice-cert\") pod \"glance-operator-controller-manager-67c6d4bf5f-ngkxb\" (UID: \"94b82631-9d81-4926-8677-1d9addae8808\") " pod="openstack-operators/glance-operator-controller-manager-67c6d4bf5f-ngkxb" Oct 04 03:26:49 crc kubenswrapper[4742]: I1004 03:26:49.433137 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kc527\" (UniqueName: \"kubernetes.io/projected/94b82631-9d81-4926-8677-1d9addae8808-kube-api-access-kc527\") pod \"glance-operator-controller-manager-67c6d4bf5f-ngkxb\" (UID: \"94b82631-9d81-4926-8677-1d9addae8808\") " pod="openstack-operators/glance-operator-controller-manager-67c6d4bf5f-ngkxb" Oct 04 03:26:49 crc kubenswrapper[4742]: I1004 03:26:49.433250 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/94b82631-9d81-4926-8677-1d9addae8808-webhook-cert\") pod \"glance-operator-controller-manager-67c6d4bf5f-ngkxb\" (UID: \"94b82631-9d81-4926-8677-1d9addae8808\") " pod="openstack-operators/glance-operator-controller-manager-67c6d4bf5f-ngkxb" Oct 04 03:26:49 crc kubenswrapper[4742]: I1004 03:26:49.433335 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/94b82631-9d81-4926-8677-1d9addae8808-apiservice-cert\") pod \"glance-operator-controller-manager-67c6d4bf5f-ngkxb\" (UID: \"94b82631-9d81-4926-8677-1d9addae8808\") " pod="openstack-operators/glance-operator-controller-manager-67c6d4bf5f-ngkxb" Oct 04 03:26:49 crc kubenswrapper[4742]: I1004 03:26:49.449012 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/94b82631-9d81-4926-8677-1d9addae8808-apiservice-cert\") pod \"glance-operator-controller-manager-67c6d4bf5f-ngkxb\" (UID: \"94b82631-9d81-4926-8677-1d9addae8808\") " pod="openstack-operators/glance-operator-controller-manager-67c6d4bf5f-ngkxb" Oct 04 03:26:49 crc kubenswrapper[4742]: I1004 03:26:49.449124 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/94b82631-9d81-4926-8677-1d9addae8808-webhook-cert\") pod \"glance-operator-controller-manager-67c6d4bf5f-ngkxb\" (UID: \"94b82631-9d81-4926-8677-1d9addae8808\") " pod="openstack-operators/glance-operator-controller-manager-67c6d4bf5f-ngkxb" Oct 04 03:26:49 crc kubenswrapper[4742]: I1004 03:26:49.456315 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kc527\" (UniqueName: \"kubernetes.io/projected/94b82631-9d81-4926-8677-1d9addae8808-kube-api-access-kc527\") pod \"glance-operator-controller-manager-67c6d4bf5f-ngkxb\" (UID: \"94b82631-9d81-4926-8677-1d9addae8808\") " pod="openstack-operators/glance-operator-controller-manager-67c6d4bf5f-ngkxb" Oct 04 03:26:49 crc kubenswrapper[4742]: I1004 03:26:49.592160 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-67c6d4bf5f-ngkxb" Oct 04 03:26:50 crc kubenswrapper[4742]: I1004 03:26:50.063204 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-67c6d4bf5f-ngkxb"] Oct 04 03:26:50 crc kubenswrapper[4742]: I1004 03:26:50.335662 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/swift-proxy-59cb459c9f-hs28w" Oct 04 03:26:50 crc kubenswrapper[4742]: I1004 03:26:50.337429 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/swift-proxy-59cb459c9f-hs28w" Oct 04 03:26:50 crc kubenswrapper[4742]: I1004 03:26:50.485406 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-67c6d4bf5f-ngkxb" event={"ID":"94b82631-9d81-4926-8677-1d9addae8808","Type":"ContainerStarted","Data":"f7f0d7905b45d4a6c83b55ffb1352cbf919ffb147adce0493a70318f99198cb4"} Oct 04 03:26:52 crc kubenswrapper[4742]: I1004 03:26:52.514052 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-67c6d4bf5f-ngkxb" event={"ID":"94b82631-9d81-4926-8677-1d9addae8808","Type":"ContainerStarted","Data":"9fc3ba208398feec50e0c043bc0bd5ee548b71407c6fca1f331951a1cd21fe26"} Oct 04 03:26:52 crc kubenswrapper[4742]: I1004 03:26:52.514944 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-67c6d4bf5f-ngkxb" Oct 04 03:26:52 crc kubenswrapper[4742]: I1004 03:26:52.514959 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-67c6d4bf5f-ngkxb" event={"ID":"94b82631-9d81-4926-8677-1d9addae8808","Type":"ContainerStarted","Data":"015a11cffa2e0ea862e75e9daad6fabbd219ea52e8dfe787c2571f6dbc4c01bc"} Oct 04 03:26:52 crc kubenswrapper[4742]: I1004 03:26:52.544166 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-67c6d4bf5f-ngkxb" podStartSLOduration=1.356987922 podStartE2EDuration="3.544141364s" podCreationTimestamp="2025-10-04 03:26:49 +0000 UTC" firstStartedPulling="2025-10-04 03:26:50.065646683 +0000 UTC m=+975.997983859" lastFinishedPulling="2025-10-04 03:26:52.252800115 +0000 UTC m=+978.185137301" observedRunningTime="2025-10-04 03:26:52.540324706 +0000 UTC m=+978.472661902" watchObservedRunningTime="2025-10-04 03:26:52.544141364 +0000 UTC m=+978.476478580" Oct 04 03:26:59 crc kubenswrapper[4742]: I1004 03:26:59.598040 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-67c6d4bf5f-ngkxb" Oct 04 03:27:02 crc kubenswrapper[4742]: I1004 03:27:02.302615 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstackclient"] Oct 04 03:27:02 crc kubenswrapper[4742]: I1004 03:27:02.304036 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Oct 04 03:27:02 crc kubenswrapper[4742]: I1004 03:27:02.306265 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"openstack-config-secret" Oct 04 03:27:02 crc kubenswrapper[4742]: I1004 03:27:02.306895 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts-9db6gc427h" Oct 04 03:27:02 crc kubenswrapper[4742]: I1004 03:27:02.307086 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"default-dockercfg-vcgns" Oct 04 03:27:02 crc kubenswrapper[4742]: I1004 03:27:02.310767 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config" Oct 04 03:27:02 crc kubenswrapper[4742]: I1004 03:27:02.321142 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Oct 04 03:27:02 crc kubenswrapper[4742]: I1004 03:27:02.401708 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-mgmlq"] Oct 04 03:27:02 crc kubenswrapper[4742]: I1004 03:27:02.408144 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-mgmlq" Oct 04 03:27:02 crc kubenswrapper[4742]: I1004 03:27:02.410398 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-mgmlq"] Oct 04 03:27:02 crc kubenswrapper[4742]: I1004 03:27:02.464315 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47zlz\" (UniqueName: \"kubernetes.io/projected/8d7e48d9-5aba-44f1-a31f-fb353d8e034d-kube-api-access-47zlz\") pod \"openstackclient\" (UID: \"8d7e48d9-5aba-44f1-a31f-fb353d8e034d\") " pod="glance-kuttl-tests/openstackclient" Oct 04 03:27:02 crc kubenswrapper[4742]: I1004 03:27:02.464375 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8d7e48d9-5aba-44f1-a31f-fb353d8e034d-openstack-config\") pod \"openstackclient\" (UID: \"8d7e48d9-5aba-44f1-a31f-fb353d8e034d\") " pod="glance-kuttl-tests/openstackclient" Oct 04 03:27:02 crc kubenswrapper[4742]: I1004 03:27:02.464438 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/8d7e48d9-5aba-44f1-a31f-fb353d8e034d-openstack-scripts\") pod \"openstackclient\" (UID: \"8d7e48d9-5aba-44f1-a31f-fb353d8e034d\") " pod="glance-kuttl-tests/openstackclient" Oct 04 03:27:02 crc kubenswrapper[4742]: I1004 03:27:02.466611 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8d7e48d9-5aba-44f1-a31f-fb353d8e034d-openstack-config-secret\") pod \"openstackclient\" (UID: \"8d7e48d9-5aba-44f1-a31f-fb353d8e034d\") " pod="glance-kuttl-tests/openstackclient" Oct 04 03:27:02 crc kubenswrapper[4742]: I1004 03:27:02.567797 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92s89\" (UniqueName: \"kubernetes.io/projected/8fdc4a83-9c85-4de1-b2ba-6bfd995d41c8-kube-api-access-92s89\") pod \"glance-db-create-mgmlq\" (UID: \"8fdc4a83-9c85-4de1-b2ba-6bfd995d41c8\") " pod="glance-kuttl-tests/glance-db-create-mgmlq" Oct 04 03:27:02 crc kubenswrapper[4742]: I1004 03:27:02.567905 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47zlz\" (UniqueName: \"kubernetes.io/projected/8d7e48d9-5aba-44f1-a31f-fb353d8e034d-kube-api-access-47zlz\") pod \"openstackclient\" (UID: \"8d7e48d9-5aba-44f1-a31f-fb353d8e034d\") " pod="glance-kuttl-tests/openstackclient" Oct 04 03:27:02 crc kubenswrapper[4742]: I1004 03:27:02.567954 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8d7e48d9-5aba-44f1-a31f-fb353d8e034d-openstack-config\") pod \"openstackclient\" (UID: \"8d7e48d9-5aba-44f1-a31f-fb353d8e034d\") " pod="glance-kuttl-tests/openstackclient" Oct 04 03:27:02 crc kubenswrapper[4742]: I1004 03:27:02.568019 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/8d7e48d9-5aba-44f1-a31f-fb353d8e034d-openstack-scripts\") pod \"openstackclient\" (UID: \"8d7e48d9-5aba-44f1-a31f-fb353d8e034d\") " pod="glance-kuttl-tests/openstackclient" Oct 04 03:27:02 crc kubenswrapper[4742]: I1004 03:27:02.568051 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8d7e48d9-5aba-44f1-a31f-fb353d8e034d-openstack-config-secret\") pod \"openstackclient\" (UID: \"8d7e48d9-5aba-44f1-a31f-fb353d8e034d\") " pod="glance-kuttl-tests/openstackclient" Oct 04 03:27:02 crc kubenswrapper[4742]: I1004 03:27:02.569441 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/8d7e48d9-5aba-44f1-a31f-fb353d8e034d-openstack-scripts\") pod \"openstackclient\" (UID: \"8d7e48d9-5aba-44f1-a31f-fb353d8e034d\") " pod="glance-kuttl-tests/openstackclient" Oct 04 03:27:02 crc kubenswrapper[4742]: I1004 03:27:02.569971 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8d7e48d9-5aba-44f1-a31f-fb353d8e034d-openstack-config\") pod \"openstackclient\" (UID: \"8d7e48d9-5aba-44f1-a31f-fb353d8e034d\") " pod="glance-kuttl-tests/openstackclient" Oct 04 03:27:02 crc kubenswrapper[4742]: I1004 03:27:02.588523 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8d7e48d9-5aba-44f1-a31f-fb353d8e034d-openstack-config-secret\") pod \"openstackclient\" (UID: \"8d7e48d9-5aba-44f1-a31f-fb353d8e034d\") " pod="glance-kuttl-tests/openstackclient" Oct 04 03:27:02 crc kubenswrapper[4742]: I1004 03:27:02.616327 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47zlz\" (UniqueName: \"kubernetes.io/projected/8d7e48d9-5aba-44f1-a31f-fb353d8e034d-kube-api-access-47zlz\") pod \"openstackclient\" (UID: \"8d7e48d9-5aba-44f1-a31f-fb353d8e034d\") " pod="glance-kuttl-tests/openstackclient" Oct 04 03:27:02 crc kubenswrapper[4742]: I1004 03:27:02.667156 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Oct 04 03:27:02 crc kubenswrapper[4742]: I1004 03:27:02.668689 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92s89\" (UniqueName: \"kubernetes.io/projected/8fdc4a83-9c85-4de1-b2ba-6bfd995d41c8-kube-api-access-92s89\") pod \"glance-db-create-mgmlq\" (UID: \"8fdc4a83-9c85-4de1-b2ba-6bfd995d41c8\") " pod="glance-kuttl-tests/glance-db-create-mgmlq" Oct 04 03:27:02 crc kubenswrapper[4742]: I1004 03:27:02.696886 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92s89\" (UniqueName: \"kubernetes.io/projected/8fdc4a83-9c85-4de1-b2ba-6bfd995d41c8-kube-api-access-92s89\") pod \"glance-db-create-mgmlq\" (UID: \"8fdc4a83-9c85-4de1-b2ba-6bfd995d41c8\") " pod="glance-kuttl-tests/glance-db-create-mgmlq" Oct 04 03:27:02 crc kubenswrapper[4742]: I1004 03:27:02.727498 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-mgmlq" Oct 04 03:27:03 crc kubenswrapper[4742]: I1004 03:27:03.097431 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Oct 04 03:27:03 crc kubenswrapper[4742]: I1004 03:27:03.240550 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-mgmlq"] Oct 04 03:27:03 crc kubenswrapper[4742]: W1004 03:27:03.240679 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8fdc4a83_9c85_4de1_b2ba_6bfd995d41c8.slice/crio-385ee01792213ea8b67a839aafcaa3af637ddaa144d9be8b7fde0e1ef40f97ce WatchSource:0}: Error finding container 385ee01792213ea8b67a839aafcaa3af637ddaa144d9be8b7fde0e1ef40f97ce: Status 404 returned error can't find the container with id 385ee01792213ea8b67a839aafcaa3af637ddaa144d9be8b7fde0e1ef40f97ce Oct 04 03:27:03 crc kubenswrapper[4742]: I1004 03:27:03.629831 4742 generic.go:334] "Generic (PLEG): container finished" podID="8fdc4a83-9c85-4de1-b2ba-6bfd995d41c8" containerID="b37287f10fe9c704fc8c003dbf0705cce72ee37d28e6ab44f4abe12d8af869f8" exitCode=0 Oct 04 03:27:03 crc kubenswrapper[4742]: I1004 03:27:03.629934 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-mgmlq" event={"ID":"8fdc4a83-9c85-4de1-b2ba-6bfd995d41c8","Type":"ContainerDied","Data":"b37287f10fe9c704fc8c003dbf0705cce72ee37d28e6ab44f4abe12d8af869f8"} Oct 04 03:27:03 crc kubenswrapper[4742]: I1004 03:27:03.630410 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-mgmlq" event={"ID":"8fdc4a83-9c85-4de1-b2ba-6bfd995d41c8","Type":"ContainerStarted","Data":"385ee01792213ea8b67a839aafcaa3af637ddaa144d9be8b7fde0e1ef40f97ce"} Oct 04 03:27:03 crc kubenswrapper[4742]: I1004 03:27:03.631300 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"8d7e48d9-5aba-44f1-a31f-fb353d8e034d","Type":"ContainerStarted","Data":"720e3c018441a8e99cebcf93f5cbcace6a3cd9433e8c20fb4afbfef51970794f"} Oct 04 03:27:05 crc kubenswrapper[4742]: I1004 03:27:05.512564 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-mgmlq" Oct 04 03:27:05 crc kubenswrapper[4742]: I1004 03:27:05.617730 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92s89\" (UniqueName: \"kubernetes.io/projected/8fdc4a83-9c85-4de1-b2ba-6bfd995d41c8-kube-api-access-92s89\") pod \"8fdc4a83-9c85-4de1-b2ba-6bfd995d41c8\" (UID: \"8fdc4a83-9c85-4de1-b2ba-6bfd995d41c8\") " Oct 04 03:27:05 crc kubenswrapper[4742]: I1004 03:27:05.626518 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fdc4a83-9c85-4de1-b2ba-6bfd995d41c8-kube-api-access-92s89" (OuterVolumeSpecName: "kube-api-access-92s89") pod "8fdc4a83-9c85-4de1-b2ba-6bfd995d41c8" (UID: "8fdc4a83-9c85-4de1-b2ba-6bfd995d41c8"). InnerVolumeSpecName "kube-api-access-92s89". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:27:05 crc kubenswrapper[4742]: I1004 03:27:05.658115 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-mgmlq" event={"ID":"8fdc4a83-9c85-4de1-b2ba-6bfd995d41c8","Type":"ContainerDied","Data":"385ee01792213ea8b67a839aafcaa3af637ddaa144d9be8b7fde0e1ef40f97ce"} Oct 04 03:27:05 crc kubenswrapper[4742]: I1004 03:27:05.658899 4742 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="385ee01792213ea8b67a839aafcaa3af637ddaa144d9be8b7fde0e1ef40f97ce" Oct 04 03:27:05 crc kubenswrapper[4742]: I1004 03:27:05.658247 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-mgmlq" Oct 04 03:27:05 crc kubenswrapper[4742]: I1004 03:27:05.719776 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92s89\" (UniqueName: \"kubernetes.io/projected/8fdc4a83-9c85-4de1-b2ba-6bfd995d41c8-kube-api-access-92s89\") on node \"crc\" DevicePath \"\"" Oct 04 03:27:11 crc kubenswrapper[4742]: I1004 03:27:11.535251 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7e8086e1-26ed-40f6-a3f7-861ec4d232f9-etc-swift\") pod \"swift-storage-0\" (UID: \"7e8086e1-26ed-40f6-a3f7-861ec4d232f9\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 03:27:11 crc kubenswrapper[4742]: I1004 03:27:11.548386 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7e8086e1-26ed-40f6-a3f7-861ec4d232f9-etc-swift\") pod \"swift-storage-0\" (UID: \"7e8086e1-26ed-40f6-a3f7-861ec4d232f9\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 03:27:11 crc kubenswrapper[4742]: I1004 03:27:11.812663 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Oct 04 03:27:12 crc kubenswrapper[4742]: I1004 03:27:12.397915 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-52a1-account-create-8zdhh"] Oct 04 03:27:12 crc kubenswrapper[4742]: E1004 03:27:12.398849 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fdc4a83-9c85-4de1-b2ba-6bfd995d41c8" containerName="mariadb-database-create" Oct 04 03:27:12 crc kubenswrapper[4742]: I1004 03:27:12.398881 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fdc4a83-9c85-4de1-b2ba-6bfd995d41c8" containerName="mariadb-database-create" Oct 04 03:27:12 crc kubenswrapper[4742]: I1004 03:27:12.399044 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fdc4a83-9c85-4de1-b2ba-6bfd995d41c8" containerName="mariadb-database-create" Oct 04 03:27:12 crc kubenswrapper[4742]: I1004 03:27:12.399772 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-52a1-account-create-8zdhh" Oct 04 03:27:12 crc kubenswrapper[4742]: I1004 03:27:12.402483 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Oct 04 03:27:12 crc kubenswrapper[4742]: I1004 03:27:12.425963 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-52a1-account-create-8zdhh"] Oct 04 03:27:12 crc kubenswrapper[4742]: I1004 03:27:12.545286 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Oct 04 03:27:12 crc kubenswrapper[4742]: I1004 03:27:12.553634 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vslf\" (UniqueName: \"kubernetes.io/projected/8e2c0638-a70e-4d1f-9cb1-ea9d43f058b4-kube-api-access-6vslf\") pod \"glance-52a1-account-create-8zdhh\" (UID: \"8e2c0638-a70e-4d1f-9cb1-ea9d43f058b4\") " pod="glance-kuttl-tests/glance-52a1-account-create-8zdhh" Oct 04 03:27:12 crc kubenswrapper[4742]: I1004 03:27:12.655077 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vslf\" (UniqueName: \"kubernetes.io/projected/8e2c0638-a70e-4d1f-9cb1-ea9d43f058b4-kube-api-access-6vslf\") pod \"glance-52a1-account-create-8zdhh\" (UID: \"8e2c0638-a70e-4d1f-9cb1-ea9d43f058b4\") " pod="glance-kuttl-tests/glance-52a1-account-create-8zdhh" Oct 04 03:27:12 crc kubenswrapper[4742]: I1004 03:27:12.677356 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vslf\" (UniqueName: \"kubernetes.io/projected/8e2c0638-a70e-4d1f-9cb1-ea9d43f058b4-kube-api-access-6vslf\") pod \"glance-52a1-account-create-8zdhh\" (UID: \"8e2c0638-a70e-4d1f-9cb1-ea9d43f058b4\") " pod="glance-kuttl-tests/glance-52a1-account-create-8zdhh" Oct 04 03:27:12 crc kubenswrapper[4742]: I1004 03:27:12.718286 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-52a1-account-create-8zdhh" Oct 04 03:27:12 crc kubenswrapper[4742]: I1004 03:27:12.729053 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"8d7e48d9-5aba-44f1-a31f-fb353d8e034d","Type":"ContainerStarted","Data":"2c5d6c5186fcc36d6be9f9b3c45e55eb3c383e9f0d1bae92257f78ab051b135f"} Oct 04 03:27:12 crc kubenswrapper[4742]: I1004 03:27:12.731875 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"7e8086e1-26ed-40f6-a3f7-861ec4d232f9","Type":"ContainerStarted","Data":"9c2623ab13e859d373ad36d54680e9d1f149c173d63770ef8402da4c77913eef"} Oct 04 03:27:12 crc kubenswrapper[4742]: I1004 03:27:12.753541 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstackclient" podStartSLOduration=1.639219562 podStartE2EDuration="10.753513949s" podCreationTimestamp="2025-10-04 03:27:02 +0000 UTC" firstStartedPulling="2025-10-04 03:27:03.106766398 +0000 UTC m=+989.039103604" lastFinishedPulling="2025-10-04 03:27:12.221060815 +0000 UTC m=+998.153397991" observedRunningTime="2025-10-04 03:27:12.747592377 +0000 UTC m=+998.679929553" watchObservedRunningTime="2025-10-04 03:27:12.753513949 +0000 UTC m=+998.685851145" Oct 04 03:27:13 crc kubenswrapper[4742]: I1004 03:27:13.167514 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-52a1-account-create-8zdhh"] Oct 04 03:27:13 crc kubenswrapper[4742]: W1004 03:27:13.175372 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8e2c0638_a70e_4d1f_9cb1_ea9d43f058b4.slice/crio-cb43a20d8b8617207a2de963c5648a4cf0698fff019cbd4ffa9ba95090bc8c26 WatchSource:0}: Error finding container cb43a20d8b8617207a2de963c5648a4cf0698fff019cbd4ffa9ba95090bc8c26: Status 404 returned error can't find the container with id cb43a20d8b8617207a2de963c5648a4cf0698fff019cbd4ffa9ba95090bc8c26 Oct 04 03:27:13 crc kubenswrapper[4742]: I1004 03:27:13.742526 4742 generic.go:334] "Generic (PLEG): container finished" podID="8e2c0638-a70e-4d1f-9cb1-ea9d43f058b4" containerID="03e6a2189290bf48fb3bced0fbf093a73e0ee75e232c27ce5b02bc4b67d08c7d" exitCode=0 Oct 04 03:27:13 crc kubenswrapper[4742]: I1004 03:27:13.742637 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-52a1-account-create-8zdhh" event={"ID":"8e2c0638-a70e-4d1f-9cb1-ea9d43f058b4","Type":"ContainerDied","Data":"03e6a2189290bf48fb3bced0fbf093a73e0ee75e232c27ce5b02bc4b67d08c7d"} Oct 04 03:27:13 crc kubenswrapper[4742]: I1004 03:27:13.742721 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-52a1-account-create-8zdhh" event={"ID":"8e2c0638-a70e-4d1f-9cb1-ea9d43f058b4","Type":"ContainerStarted","Data":"cb43a20d8b8617207a2de963c5648a4cf0698fff019cbd4ffa9ba95090bc8c26"} Oct 04 03:27:14 crc kubenswrapper[4742]: I1004 03:27:14.883066 4742 patch_prober.go:28] interesting pod/machine-config-daemon-qdt7v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:27:14 crc kubenswrapper[4742]: I1004 03:27:14.883120 4742 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:27:15 crc kubenswrapper[4742]: I1004 03:27:15.112853 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-52a1-account-create-8zdhh" Oct 04 03:27:15 crc kubenswrapper[4742]: I1004 03:27:15.202190 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vslf\" (UniqueName: \"kubernetes.io/projected/8e2c0638-a70e-4d1f-9cb1-ea9d43f058b4-kube-api-access-6vslf\") pod \"8e2c0638-a70e-4d1f-9cb1-ea9d43f058b4\" (UID: \"8e2c0638-a70e-4d1f-9cb1-ea9d43f058b4\") " Oct 04 03:27:15 crc kubenswrapper[4742]: I1004 03:27:15.211852 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e2c0638-a70e-4d1f-9cb1-ea9d43f058b4-kube-api-access-6vslf" (OuterVolumeSpecName: "kube-api-access-6vslf") pod "8e2c0638-a70e-4d1f-9cb1-ea9d43f058b4" (UID: "8e2c0638-a70e-4d1f-9cb1-ea9d43f058b4"). InnerVolumeSpecName "kube-api-access-6vslf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:27:15 crc kubenswrapper[4742]: I1004 03:27:15.304801 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vslf\" (UniqueName: \"kubernetes.io/projected/8e2c0638-a70e-4d1f-9cb1-ea9d43f058b4-kube-api-access-6vslf\") on node \"crc\" DevicePath \"\"" Oct 04 03:27:15 crc kubenswrapper[4742]: I1004 03:27:15.763989 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-52a1-account-create-8zdhh" event={"ID":"8e2c0638-a70e-4d1f-9cb1-ea9d43f058b4","Type":"ContainerDied","Data":"cb43a20d8b8617207a2de963c5648a4cf0698fff019cbd4ffa9ba95090bc8c26"} Oct 04 03:27:15 crc kubenswrapper[4742]: I1004 03:27:15.764350 4742 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb43a20d8b8617207a2de963c5648a4cf0698fff019cbd4ffa9ba95090bc8c26" Oct 04 03:27:15 crc kubenswrapper[4742]: I1004 03:27:15.764465 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-52a1-account-create-8zdhh" Oct 04 03:27:15 crc kubenswrapper[4742]: I1004 03:27:15.766967 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"7e8086e1-26ed-40f6-a3f7-861ec4d232f9","Type":"ContainerStarted","Data":"2b02ae78c2f30b192ddc8d05d65593af53dece42e528b6d5561fc78507a346b6"} Oct 04 03:27:15 crc kubenswrapper[4742]: I1004 03:27:15.766992 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"7e8086e1-26ed-40f6-a3f7-861ec4d232f9","Type":"ContainerStarted","Data":"ef31987f0e9c68c842a60357c515e51e201e635061bb9b3ca909ae49e284369c"} Oct 04 03:27:17 crc kubenswrapper[4742]: I1004 03:27:17.471195 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-2dq5q"] Oct 04 03:27:17 crc kubenswrapper[4742]: E1004 03:27:17.472389 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e2c0638-a70e-4d1f-9cb1-ea9d43f058b4" containerName="mariadb-account-create" Oct 04 03:27:17 crc kubenswrapper[4742]: I1004 03:27:17.472433 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e2c0638-a70e-4d1f-9cb1-ea9d43f058b4" containerName="mariadb-account-create" Oct 04 03:27:17 crc kubenswrapper[4742]: I1004 03:27:17.473003 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e2c0638-a70e-4d1f-9cb1-ea9d43f058b4" containerName="mariadb-account-create" Oct 04 03:27:17 crc kubenswrapper[4742]: I1004 03:27:17.473858 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-2dq5q" Oct 04 03:27:17 crc kubenswrapper[4742]: I1004 03:27:17.477126 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-tjlkz" Oct 04 03:27:17 crc kubenswrapper[4742]: I1004 03:27:17.478611 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Oct 04 03:27:17 crc kubenswrapper[4742]: I1004 03:27:17.486510 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-2dq5q"] Oct 04 03:27:17 crc kubenswrapper[4742]: I1004 03:27:17.540347 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6-config-data\") pod \"glance-db-sync-2dq5q\" (UID: \"b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6\") " pod="glance-kuttl-tests/glance-db-sync-2dq5q" Oct 04 03:27:17 crc kubenswrapper[4742]: I1004 03:27:17.540903 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmxlf\" (UniqueName: \"kubernetes.io/projected/b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6-kube-api-access-qmxlf\") pod \"glance-db-sync-2dq5q\" (UID: \"b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6\") " pod="glance-kuttl-tests/glance-db-sync-2dq5q" Oct 04 03:27:17 crc kubenswrapper[4742]: I1004 03:27:17.541172 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6-db-sync-config-data\") pod \"glance-db-sync-2dq5q\" (UID: \"b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6\") " pod="glance-kuttl-tests/glance-db-sync-2dq5q" Oct 04 03:27:17 crc kubenswrapper[4742]: I1004 03:27:17.643401 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmxlf\" (UniqueName: \"kubernetes.io/projected/b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6-kube-api-access-qmxlf\") pod \"glance-db-sync-2dq5q\" (UID: \"b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6\") " pod="glance-kuttl-tests/glance-db-sync-2dq5q" Oct 04 03:27:17 crc kubenswrapper[4742]: I1004 03:27:17.643958 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6-db-sync-config-data\") pod \"glance-db-sync-2dq5q\" (UID: \"b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6\") " pod="glance-kuttl-tests/glance-db-sync-2dq5q" Oct 04 03:27:17 crc kubenswrapper[4742]: I1004 03:27:17.644199 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6-config-data\") pod \"glance-db-sync-2dq5q\" (UID: \"b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6\") " pod="glance-kuttl-tests/glance-db-sync-2dq5q" Oct 04 03:27:17 crc kubenswrapper[4742]: I1004 03:27:17.663121 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6-config-data\") pod \"glance-db-sync-2dq5q\" (UID: \"b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6\") " pod="glance-kuttl-tests/glance-db-sync-2dq5q" Oct 04 03:27:17 crc kubenswrapper[4742]: I1004 03:27:17.663584 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6-db-sync-config-data\") pod \"glance-db-sync-2dq5q\" (UID: \"b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6\") " pod="glance-kuttl-tests/glance-db-sync-2dq5q" Oct 04 03:27:17 crc kubenswrapper[4742]: I1004 03:27:17.666925 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmxlf\" (UniqueName: \"kubernetes.io/projected/b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6-kube-api-access-qmxlf\") pod \"glance-db-sync-2dq5q\" (UID: \"b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6\") " pod="glance-kuttl-tests/glance-db-sync-2dq5q" Oct 04 03:27:17 crc kubenswrapper[4742]: I1004 03:27:17.786985 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"7e8086e1-26ed-40f6-a3f7-861ec4d232f9","Type":"ContainerStarted","Data":"628c9532ac797b49a9b0e24dd58d870e7f68f060a5b59c949929b831cb427520"} Oct 04 03:27:17 crc kubenswrapper[4742]: I1004 03:27:17.804743 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-2dq5q" Oct 04 03:27:18 crc kubenswrapper[4742]: I1004 03:27:18.283828 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-2dq5q"] Oct 04 03:27:18 crc kubenswrapper[4742]: I1004 03:27:18.797824 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-2dq5q" event={"ID":"b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6","Type":"ContainerStarted","Data":"9ff6f3a25bf0bd40b382e00f552634db0d06ccc073352433ec81ce6622cfacb8"} Oct 04 03:27:27 crc kubenswrapper[4742]: I1004 03:27:27.883701 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"7e8086e1-26ed-40f6-a3f7-861ec4d232f9","Type":"ContainerStarted","Data":"a9ac155e39a108cec2b7d1880b52ed4c39b029da8ab8cc76b26ed7e2330147e0"} Oct 04 03:27:42 crc kubenswrapper[4742]: I1004 03:27:42.018112 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"7e8086e1-26ed-40f6-a3f7-861ec4d232f9","Type":"ContainerStarted","Data":"059d544d3b9a74abcf11d2ea8f61dc9741ec5ec09c2e86a043fff19938c1dfee"} Oct 04 03:27:42 crc kubenswrapper[4742]: I1004 03:27:42.018845 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"7e8086e1-26ed-40f6-a3f7-861ec4d232f9","Type":"ContainerStarted","Data":"e151328e2c3c46a294b4c6767d0dfbeb64408d58e25cda60bd6328a62770efb8"} Oct 04 03:27:42 crc kubenswrapper[4742]: I1004 03:27:42.018942 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"7e8086e1-26ed-40f6-a3f7-861ec4d232f9","Type":"ContainerStarted","Data":"9c41a00a244d9da55f577a5a6d2b806c6cfd2c75ef09dc3f6eeab8cddcb3ec28"} Oct 04 03:27:42 crc kubenswrapper[4742]: I1004 03:27:42.018956 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"7e8086e1-26ed-40f6-a3f7-861ec4d232f9","Type":"ContainerStarted","Data":"ac03b211540d6406099ce47274d78c55f8647e9c42b9ecccb315767341412c39"} Oct 04 03:27:42 crc kubenswrapper[4742]: I1004 03:27:42.022688 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-2dq5q" event={"ID":"b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6","Type":"ContainerStarted","Data":"3ef20e27d6e9a853995b31d1d16cac238881afbf10f611ee3824536fc009132c"} Oct 04 03:27:42 crc kubenswrapper[4742]: I1004 03:27:42.042911 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-2dq5q" podStartSLOduration=2.326925982 podStartE2EDuration="25.042892811s" podCreationTimestamp="2025-10-04 03:27:17 +0000 UTC" firstStartedPulling="2025-10-04 03:27:18.29566992 +0000 UTC m=+1004.228007096" lastFinishedPulling="2025-10-04 03:27:41.011636709 +0000 UTC m=+1026.943973925" observedRunningTime="2025-10-04 03:27:42.041956347 +0000 UTC m=+1027.974293543" watchObservedRunningTime="2025-10-04 03:27:42.042892811 +0000 UTC m=+1027.975229987" Oct 04 03:27:44 crc kubenswrapper[4742]: I1004 03:27:44.871762 4742 patch_prober.go:28] interesting pod/machine-config-daemon-qdt7v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:27:44 crc kubenswrapper[4742]: I1004 03:27:44.872662 4742 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:27:45 crc kubenswrapper[4742]: I1004 03:27:45.050159 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"7e8086e1-26ed-40f6-a3f7-861ec4d232f9","Type":"ContainerStarted","Data":"08c62d203ddbd06fa8f148bdec6235ee29c8bc068132275fb2072d9030b40c00"} Oct 04 03:27:45 crc kubenswrapper[4742]: I1004 03:27:45.050221 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"7e8086e1-26ed-40f6-a3f7-861ec4d232f9","Type":"ContainerStarted","Data":"2d44f33093462456368c1088b5916934235fef526df35bea7e82beb18b225081"} Oct 04 03:27:45 crc kubenswrapper[4742]: I1004 03:27:45.050240 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"7e8086e1-26ed-40f6-a3f7-861ec4d232f9","Type":"ContainerStarted","Data":"9a4ff369b1a6cc36cac75647ffadae76156831ee7f5028a8c3b1283fd3c19ea4"} Oct 04 03:27:47 crc kubenswrapper[4742]: I1004 03:27:47.078709 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"7e8086e1-26ed-40f6-a3f7-861ec4d232f9","Type":"ContainerStarted","Data":"8663790a1bcccfd428a99cee48799b0b93b5c73313746091fad4f9b8de53b24d"} Oct 04 03:27:47 crc kubenswrapper[4742]: I1004 03:27:47.079255 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"7e8086e1-26ed-40f6-a3f7-861ec4d232f9","Type":"ContainerStarted","Data":"1473747ff6e185ea7ee12dc8b7214919ba1a4557cfcd1ca4dd3c5e6016210f87"} Oct 04 03:27:47 crc kubenswrapper[4742]: I1004 03:27:47.079314 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"7e8086e1-26ed-40f6-a3f7-861ec4d232f9","Type":"ContainerStarted","Data":"4f66f9280b086fd14b882cdbaa5d256d821750d473a52617709342d7218a04ba"} Oct 04 03:27:47 crc kubenswrapper[4742]: I1004 03:27:47.079335 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"7e8086e1-26ed-40f6-a3f7-861ec4d232f9","Type":"ContainerStarted","Data":"12b1918bef8eb08a8d94a98653a5eece2c29b1c488ccafd7aa4eb7242caf3e9e"} Oct 04 03:27:47 crc kubenswrapper[4742]: I1004 03:27:47.130956 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-storage-0" podStartSLOduration=69.782309783 podStartE2EDuration="1m41.130937394s" podCreationTimestamp="2025-10-04 03:26:06 +0000 UTC" firstStartedPulling="2025-10-04 03:27:12.556903969 +0000 UTC m=+998.489241145" lastFinishedPulling="2025-10-04 03:27:43.90553158 +0000 UTC m=+1029.837868756" observedRunningTime="2025-10-04 03:27:47.127060094 +0000 UTC m=+1033.059397290" watchObservedRunningTime="2025-10-04 03:27:47.130937394 +0000 UTC m=+1033.063274580" Oct 04 03:27:51 crc kubenswrapper[4742]: I1004 03:27:51.119179 4742 generic.go:334] "Generic (PLEG): container finished" podID="b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6" containerID="3ef20e27d6e9a853995b31d1d16cac238881afbf10f611ee3824536fc009132c" exitCode=0 Oct 04 03:27:51 crc kubenswrapper[4742]: I1004 03:27:51.119340 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-2dq5q" event={"ID":"b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6","Type":"ContainerDied","Data":"3ef20e27d6e9a853995b31d1d16cac238881afbf10f611ee3824536fc009132c"} Oct 04 03:27:52 crc kubenswrapper[4742]: I1004 03:27:52.429115 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-2dq5q" Oct 04 03:27:52 crc kubenswrapper[4742]: I1004 03:27:52.590263 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6-db-sync-config-data\") pod \"b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6\" (UID: \"b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6\") " Oct 04 03:27:52 crc kubenswrapper[4742]: I1004 03:27:52.590527 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmxlf\" (UniqueName: \"kubernetes.io/projected/b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6-kube-api-access-qmxlf\") pod \"b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6\" (UID: \"b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6\") " Oct 04 03:27:52 crc kubenswrapper[4742]: I1004 03:27:52.590606 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6-config-data\") pod \"b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6\" (UID: \"b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6\") " Oct 04 03:27:52 crc kubenswrapper[4742]: I1004 03:27:52.596887 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6" (UID: "b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:27:52 crc kubenswrapper[4742]: I1004 03:27:52.597577 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6-kube-api-access-qmxlf" (OuterVolumeSpecName: "kube-api-access-qmxlf") pod "b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6" (UID: "b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6"). InnerVolumeSpecName "kube-api-access-qmxlf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:27:52 crc kubenswrapper[4742]: I1004 03:27:52.633907 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6-config-data" (OuterVolumeSpecName: "config-data") pod "b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6" (UID: "b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:27:52 crc kubenswrapper[4742]: I1004 03:27:52.692928 4742 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:27:52 crc kubenswrapper[4742]: I1004 03:27:52.692962 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmxlf\" (UniqueName: \"kubernetes.io/projected/b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6-kube-api-access-qmxlf\") on node \"crc\" DevicePath \"\"" Oct 04 03:27:52 crc kubenswrapper[4742]: I1004 03:27:52.692972 4742 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:27:53 crc kubenswrapper[4742]: I1004 03:27:53.149023 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-2dq5q" event={"ID":"b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6","Type":"ContainerDied","Data":"9ff6f3a25bf0bd40b382e00f552634db0d06ccc073352433ec81ce6622cfacb8"} Oct 04 03:27:53 crc kubenswrapper[4742]: I1004 03:27:53.149515 4742 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ff6f3a25bf0bd40b382e00f552634db0d06ccc073352433ec81ce6622cfacb8" Oct 04 03:27:53 crc kubenswrapper[4742]: I1004 03:27:53.149099 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-2dq5q" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.551034 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 03:27:54 crc kubenswrapper[4742]: E1004 03:27:54.551927 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6" containerName="glance-db-sync" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.551948 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6" containerName="glance-db-sync" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.552222 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6" containerName="glance-db-sync" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.553545 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.556228 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-tjlkz" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.556553 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.556648 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.581483 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.603221 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.605251 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.615563 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.631309 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.733355 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-dev\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.733743 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db88c4d6-966a-4144-90a8-33b468f422aa-scripts\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.733920 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcgk4\" (UniqueName: \"kubernetes.io/projected/3bfe21a9-2af5-43c3-b566-3d4529e4e542-kube-api-access-zcgk4\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.734093 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-lib-modules\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.734296 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.734508 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db88c4d6-966a-4144-90a8-33b468f422aa-config-data\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.734656 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-run\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.734809 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-etc-nvme\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.734954 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/db88c4d6-966a-4144-90a8-33b468f422aa-httpd-run\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.735113 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3bfe21a9-2af5-43c3-b566-3d4529e4e542-logs\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.735300 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.735490 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjbsv\" (UniqueName: \"kubernetes.io/projected/db88c4d6-966a-4144-90a8-33b468f422aa-kube-api-access-jjbsv\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.735667 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bfe21a9-2af5-43c3-b566-3d4529e4e542-config-data\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.735860 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-etc-nvme\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.736037 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-dev\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.736200 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bfe21a9-2af5-43c3-b566-3d4529e4e542-scripts\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.736397 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.736601 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.736809 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.736974 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-sys\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.737137 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-lib-modules\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.737350 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db88c4d6-966a-4144-90a8-33b468f422aa-logs\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.737948 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.738154 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.738348 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.738567 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-run\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.738714 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3bfe21a9-2af5-43c3-b566-3d4529e4e542-httpd-run\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.738920 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-sys\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.739973 4742 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") device mount path \"/mnt/openstack/pv04\"" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.774355 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.840843 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3bfe21a9-2af5-43c3-b566-3d4529e4e542-httpd-run\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.841421 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-run\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.841476 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-sys\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.841513 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-dev\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.841551 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db88c4d6-966a-4144-90a8-33b468f422aa-scripts\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.841591 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcgk4\" (UniqueName: \"kubernetes.io/projected/3bfe21a9-2af5-43c3-b566-3d4529e4e542-kube-api-access-zcgk4\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.841611 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-run\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.841632 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-lib-modules\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.841666 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-sys\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.841698 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-lib-modules\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.841747 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-dev\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.841744 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.841803 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db88c4d6-966a-4144-90a8-33b468f422aa-config-data\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.841830 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-run\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.841893 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-etc-nvme\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.841918 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/db88c4d6-966a-4144-90a8-33b468f422aa-httpd-run\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.841942 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3bfe21a9-2af5-43c3-b566-3d4529e4e542-logs\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.841969 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.842034 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjbsv\" (UniqueName: \"kubernetes.io/projected/db88c4d6-966a-4144-90a8-33b468f422aa-kube-api-access-jjbsv\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.842098 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bfe21a9-2af5-43c3-b566-3d4529e4e542-config-data\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.842113 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-etc-nvme\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.842121 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-etc-nvme\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.842152 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-run\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.842153 4742 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") device mount path \"/mnt/openstack/pv05\"" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.842180 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-dev\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.842247 4742 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") device mount path \"/mnt/openstack/pv02\"" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.842726 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-etc-nvme\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.842156 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-dev\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.843301 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/db88c4d6-966a-4144-90a8-33b468f422aa-httpd-run\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.843304 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bfe21a9-2af5-43c3-b566-3d4529e4e542-scripts\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.843394 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.843459 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.843546 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.843579 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-sys\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.843608 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-lib-modules\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.843631 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db88c4d6-966a-4144-90a8-33b468f422aa-logs\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.843676 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.843747 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.843773 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3bfe21a9-2af5-43c3-b566-3d4529e4e542-logs\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.843807 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.843911 4742 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.843914 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.843925 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.843954 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-lib-modules\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.843964 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.843976 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-sys\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.844206 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db88c4d6-966a-4144-90a8-33b468f422aa-logs\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.845248 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3bfe21a9-2af5-43c3-b566-3d4529e4e542-httpd-run\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.847810 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db88c4d6-966a-4144-90a8-33b468f422aa-scripts\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.848169 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bfe21a9-2af5-43c3-b566-3d4529e4e542-config-data\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.848293 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bfe21a9-2af5-43c3-b566-3d4529e4e542-scripts\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.849644 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db88c4d6-966a-4144-90a8-33b468f422aa-config-data\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.870878 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcgk4\" (UniqueName: \"kubernetes.io/projected/3bfe21a9-2af5-43c3-b566-3d4529e4e542-kube-api-access-zcgk4\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.876776 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjbsv\" (UniqueName: \"kubernetes.io/projected/db88c4d6-966a-4144-90a8-33b468f422aa-kube-api-access-jjbsv\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.885836 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-1\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.899515 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.904293 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.925169 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:54 crc kubenswrapper[4742]: I1004 03:27:54.999006 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 04 03:27:55 crc kubenswrapper[4742]: I1004 03:27:55.173182 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:27:55 crc kubenswrapper[4742]: I1004 03:27:55.206727 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 04 03:27:55 crc kubenswrapper[4742]: I1004 03:27:55.662920 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.175772 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"db88c4d6-966a-4144-90a8-33b468f422aa","Type":"ContainerStarted","Data":"4e7eafaddf997c85b637881517f7ff64e36fa44aff7a02665ed2cb33e59ba606"} Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.176470 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"db88c4d6-966a-4144-90a8-33b468f422aa","Type":"ContainerStarted","Data":"cad70c7c3e4c0335aeb00f9b6533f96dbf1092bd9fe6818171b995731586a710"} Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.176538 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"db88c4d6-966a-4144-90a8-33b468f422aa","Type":"ContainerStarted","Data":"f509931cffc450cb914a27a6798487fd6de77d4635949c66637149ed385aaefb"} Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.183582 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"3bfe21a9-2af5-43c3-b566-3d4529e4e542","Type":"ContainerStarted","Data":"056538ea7a46a40e49862c57f15e3f6f397d25e6f266d745ad481f8f01591a59"} Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.183660 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"3bfe21a9-2af5-43c3-b566-3d4529e4e542","Type":"ContainerStarted","Data":"febabb48d5b2e9d432331c362fa09f39c2527d7c461e1f057cb5d3a282320cf6"} Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.183678 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"3bfe21a9-2af5-43c3-b566-3d4529e4e542","Type":"ContainerStarted","Data":"04f9caf6893c17cc51475e5e958b7ca204eb72755c6fb2794ba4f329f1dd85fa"} Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.183904 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="3bfe21a9-2af5-43c3-b566-3d4529e4e542" containerName="glance-log" containerID="cri-o://febabb48d5b2e9d432331c362fa09f39c2527d7c461e1f057cb5d3a282320cf6" gracePeriod=30 Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.184708 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="3bfe21a9-2af5-43c3-b566-3d4529e4e542" containerName="glance-httpd" containerID="cri-o://056538ea7a46a40e49862c57f15e3f6f397d25e6f266d745ad481f8f01591a59" gracePeriod=30 Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.243406 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-1" podStartSLOduration=3.24338062 podStartE2EDuration="3.24338062s" podCreationTimestamp="2025-10-04 03:27:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:27:56.238854304 +0000 UTC m=+1042.171191520" watchObservedRunningTime="2025-10-04 03:27:56.24338062 +0000 UTC m=+1042.175717806" Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.245393 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=3.245381311 podStartE2EDuration="3.245381311s" podCreationTimestamp="2025-10-04 03:27:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:27:56.215103974 +0000 UTC m=+1042.147441160" watchObservedRunningTime="2025-10-04 03:27:56.245381311 +0000 UTC m=+1042.177718497" Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.579145 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.675347 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-etc-nvme\") pod \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.675402 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-etc-iscsi\") pod \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.675431 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.675454 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "3bfe21a9-2af5-43c3-b566-3d4529e4e542" (UID: "3bfe21a9-2af5-43c3-b566-3d4529e4e542"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.675502 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-lib-modules\") pod \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.675541 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bfe21a9-2af5-43c3-b566-3d4529e4e542-config-data\") pod \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.675573 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.675606 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3bfe21a9-2af5-43c3-b566-3d4529e4e542-logs\") pod \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.675639 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-var-locks-brick\") pod \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.675648 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "3bfe21a9-2af5-43c3-b566-3d4529e4e542" (UID: "3bfe21a9-2af5-43c3-b566-3d4529e4e542"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.675685 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3bfe21a9-2af5-43c3-b566-3d4529e4e542-httpd-run\") pod \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.675756 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "3bfe21a9-2af5-43c3-b566-3d4529e4e542" (UID: "3bfe21a9-2af5-43c3-b566-3d4529e4e542"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.675809 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-run\") pod \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.675867 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "3bfe21a9-2af5-43c3-b566-3d4529e4e542" (UID: "3bfe21a9-2af5-43c3-b566-3d4529e4e542"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.675906 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcgk4\" (UniqueName: \"kubernetes.io/projected/3bfe21a9-2af5-43c3-b566-3d4529e4e542-kube-api-access-zcgk4\") pod \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.675927 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-run" (OuterVolumeSpecName: "run") pod "3bfe21a9-2af5-43c3-b566-3d4529e4e542" (UID: "3bfe21a9-2af5-43c3-b566-3d4529e4e542"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.675993 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bfe21a9-2af5-43c3-b566-3d4529e4e542-scripts\") pod \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.676009 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bfe21a9-2af5-43c3-b566-3d4529e4e542-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3bfe21a9-2af5-43c3-b566-3d4529e4e542" (UID: "3bfe21a9-2af5-43c3-b566-3d4529e4e542"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.676058 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-sys\") pod \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.676106 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-dev\") pod \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\" (UID: \"3bfe21a9-2af5-43c3-b566-3d4529e4e542\") " Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.676163 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-sys" (OuterVolumeSpecName: "sys") pod "3bfe21a9-2af5-43c3-b566-3d4529e4e542" (UID: "3bfe21a9-2af5-43c3-b566-3d4529e4e542"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.676231 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bfe21a9-2af5-43c3-b566-3d4529e4e542-logs" (OuterVolumeSpecName: "logs") pod "3bfe21a9-2af5-43c3-b566-3d4529e4e542" (UID: "3bfe21a9-2af5-43c3-b566-3d4529e4e542"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.676238 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-dev" (OuterVolumeSpecName: "dev") pod "3bfe21a9-2af5-43c3-b566-3d4529e4e542" (UID: "3bfe21a9-2af5-43c3-b566-3d4529e4e542"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.677190 4742 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-dev\") on node \"crc\" DevicePath \"\"" Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.677221 4742 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-etc-nvme\") on node \"crc\" DevicePath \"\"" Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.677242 4742 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-etc-iscsi\") on node \"crc\" DevicePath \"\"" Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.677259 4742 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-lib-modules\") on node \"crc\" DevicePath \"\"" Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.677304 4742 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3bfe21a9-2af5-43c3-b566-3d4529e4e542-logs\") on node \"crc\" DevicePath \"\"" Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.677321 4742 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-var-locks-brick\") on node \"crc\" DevicePath \"\"" Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.677341 4742 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3bfe21a9-2af5-43c3-b566-3d4529e4e542-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.677361 4742 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-run\") on node \"crc\" DevicePath \"\"" Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.677377 4742 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3bfe21a9-2af5-43c3-b566-3d4529e4e542-sys\") on node \"crc\" DevicePath \"\"" Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.693631 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance-cache") pod "3bfe21a9-2af5-43c3-b566-3d4529e4e542" (UID: "3bfe21a9-2af5-43c3-b566-3d4529e4e542"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.694536 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "3bfe21a9-2af5-43c3-b566-3d4529e4e542" (UID: "3bfe21a9-2af5-43c3-b566-3d4529e4e542"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.696574 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bfe21a9-2af5-43c3-b566-3d4529e4e542-scripts" (OuterVolumeSpecName: "scripts") pod "3bfe21a9-2af5-43c3-b566-3d4529e4e542" (UID: "3bfe21a9-2af5-43c3-b566-3d4529e4e542"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.730491 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bfe21a9-2af5-43c3-b566-3d4529e4e542-kube-api-access-zcgk4" (OuterVolumeSpecName: "kube-api-access-zcgk4") pod "3bfe21a9-2af5-43c3-b566-3d4529e4e542" (UID: "3bfe21a9-2af5-43c3-b566-3d4529e4e542"). InnerVolumeSpecName "kube-api-access-zcgk4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.783063 4742 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.783584 4742 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.783597 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcgk4\" (UniqueName: \"kubernetes.io/projected/3bfe21a9-2af5-43c3-b566-3d4529e4e542-kube-api-access-zcgk4\") on node \"crc\" DevicePath \"\"" Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.783609 4742 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bfe21a9-2af5-43c3-b566-3d4529e4e542-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.826502 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bfe21a9-2af5-43c3-b566-3d4529e4e542-config-data" (OuterVolumeSpecName: "config-data") pod "3bfe21a9-2af5-43c3-b566-3d4529e4e542" (UID: "3bfe21a9-2af5-43c3-b566-3d4529e4e542"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.831153 4742 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.877003 4742 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.885471 4742 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.885505 4742 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bfe21a9-2af5-43c3-b566-3d4529e4e542-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:27:56 crc kubenswrapper[4742]: I1004 03:27:56.885517 4742 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.196187 4742 generic.go:334] "Generic (PLEG): container finished" podID="3bfe21a9-2af5-43c3-b566-3d4529e4e542" containerID="056538ea7a46a40e49862c57f15e3f6f397d25e6f266d745ad481f8f01591a59" exitCode=143 Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.196256 4742 generic.go:334] "Generic (PLEG): container finished" podID="3bfe21a9-2af5-43c3-b566-3d4529e4e542" containerID="febabb48d5b2e9d432331c362fa09f39c2527d7c461e1f057cb5d3a282320cf6" exitCode=143 Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.196669 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.197216 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"3bfe21a9-2af5-43c3-b566-3d4529e4e542","Type":"ContainerDied","Data":"056538ea7a46a40e49862c57f15e3f6f397d25e6f266d745ad481f8f01591a59"} Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.197351 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"3bfe21a9-2af5-43c3-b566-3d4529e4e542","Type":"ContainerDied","Data":"febabb48d5b2e9d432331c362fa09f39c2527d7c461e1f057cb5d3a282320cf6"} Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.197382 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"3bfe21a9-2af5-43c3-b566-3d4529e4e542","Type":"ContainerDied","Data":"04f9caf6893c17cc51475e5e958b7ca204eb72755c6fb2794ba4f329f1dd85fa"} Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.197417 4742 scope.go:117] "RemoveContainer" containerID="056538ea7a46a40e49862c57f15e3f6f397d25e6f266d745ad481f8f01591a59" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.242633 4742 scope.go:117] "RemoveContainer" containerID="febabb48d5b2e9d432331c362fa09f39c2527d7c461e1f057cb5d3a282320cf6" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.270295 4742 scope.go:117] "RemoveContainer" containerID="056538ea7a46a40e49862c57f15e3f6f397d25e6f266d745ad481f8f01591a59" Oct 04 03:27:57 crc kubenswrapper[4742]: E1004 03:27:57.271206 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"056538ea7a46a40e49862c57f15e3f6f397d25e6f266d745ad481f8f01591a59\": container with ID starting with 056538ea7a46a40e49862c57f15e3f6f397d25e6f266d745ad481f8f01591a59 not found: ID does not exist" containerID="056538ea7a46a40e49862c57f15e3f6f397d25e6f266d745ad481f8f01591a59" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.271259 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"056538ea7a46a40e49862c57f15e3f6f397d25e6f266d745ad481f8f01591a59"} err="failed to get container status \"056538ea7a46a40e49862c57f15e3f6f397d25e6f266d745ad481f8f01591a59\": rpc error: code = NotFound desc = could not find container \"056538ea7a46a40e49862c57f15e3f6f397d25e6f266d745ad481f8f01591a59\": container with ID starting with 056538ea7a46a40e49862c57f15e3f6f397d25e6f266d745ad481f8f01591a59 not found: ID does not exist" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.271338 4742 scope.go:117] "RemoveContainer" containerID="febabb48d5b2e9d432331c362fa09f39c2527d7c461e1f057cb5d3a282320cf6" Oct 04 03:27:57 crc kubenswrapper[4742]: E1004 03:27:57.271888 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"febabb48d5b2e9d432331c362fa09f39c2527d7c461e1f057cb5d3a282320cf6\": container with ID starting with febabb48d5b2e9d432331c362fa09f39c2527d7c461e1f057cb5d3a282320cf6 not found: ID does not exist" containerID="febabb48d5b2e9d432331c362fa09f39c2527d7c461e1f057cb5d3a282320cf6" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.271918 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"febabb48d5b2e9d432331c362fa09f39c2527d7c461e1f057cb5d3a282320cf6"} err="failed to get container status \"febabb48d5b2e9d432331c362fa09f39c2527d7c461e1f057cb5d3a282320cf6\": rpc error: code = NotFound desc = could not find container \"febabb48d5b2e9d432331c362fa09f39c2527d7c461e1f057cb5d3a282320cf6\": container with ID starting with febabb48d5b2e9d432331c362fa09f39c2527d7c461e1f057cb5d3a282320cf6 not found: ID does not exist" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.271933 4742 scope.go:117] "RemoveContainer" containerID="056538ea7a46a40e49862c57f15e3f6f397d25e6f266d745ad481f8f01591a59" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.272210 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"056538ea7a46a40e49862c57f15e3f6f397d25e6f266d745ad481f8f01591a59"} err="failed to get container status \"056538ea7a46a40e49862c57f15e3f6f397d25e6f266d745ad481f8f01591a59\": rpc error: code = NotFound desc = could not find container \"056538ea7a46a40e49862c57f15e3f6f397d25e6f266d745ad481f8f01591a59\": container with ID starting with 056538ea7a46a40e49862c57f15e3f6f397d25e6f266d745ad481f8f01591a59 not found: ID does not exist" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.272231 4742 scope.go:117] "RemoveContainer" containerID="febabb48d5b2e9d432331c362fa09f39c2527d7c461e1f057cb5d3a282320cf6" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.274321 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"febabb48d5b2e9d432331c362fa09f39c2527d7c461e1f057cb5d3a282320cf6"} err="failed to get container status \"febabb48d5b2e9d432331c362fa09f39c2527d7c461e1f057cb5d3a282320cf6\": rpc error: code = NotFound desc = could not find container \"febabb48d5b2e9d432331c362fa09f39c2527d7c461e1f057cb5d3a282320cf6\": container with ID starting with febabb48d5b2e9d432331c362fa09f39c2527d7c461e1f057cb5d3a282320cf6 not found: ID does not exist" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.280538 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.289964 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.295570 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 04 03:27:57 crc kubenswrapper[4742]: E1004 03:27:57.295928 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bfe21a9-2af5-43c3-b566-3d4529e4e542" containerName="glance-httpd" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.295953 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bfe21a9-2af5-43c3-b566-3d4529e4e542" containerName="glance-httpd" Oct 04 03:27:57 crc kubenswrapper[4742]: E1004 03:27:57.295996 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bfe21a9-2af5-43c3-b566-3d4529e4e542" containerName="glance-log" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.296003 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bfe21a9-2af5-43c3-b566-3d4529e4e542" containerName="glance-log" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.296136 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bfe21a9-2af5-43c3-b566-3d4529e4e542" containerName="glance-httpd" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.296161 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bfe21a9-2af5-43c3-b566-3d4529e4e542" containerName="glance-log" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.296910 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.301405 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.393515 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.393564 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-run\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.393593 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.393616 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/10ce3a20-645d-4c51-9736-87104a27f12f-httpd-run\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.393826 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.393891 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10ce3a20-645d-4c51-9736-87104a27f12f-logs\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.393962 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10ce3a20-645d-4c51-9736-87104a27f12f-scripts\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.394029 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sm7nq\" (UniqueName: \"kubernetes.io/projected/10ce3a20-645d-4c51-9736-87104a27f12f-kube-api-access-sm7nq\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.394139 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.394164 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10ce3a20-645d-4c51-9736-87104a27f12f-config-data\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.394228 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-dev\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.394345 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-lib-modules\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.394478 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-sys\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.394508 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-etc-nvme\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.495505 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.495575 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/10ce3a20-645d-4c51-9736-87104a27f12f-httpd-run\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.495616 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.495637 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10ce3a20-645d-4c51-9736-87104a27f12f-logs\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.495658 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10ce3a20-645d-4c51-9736-87104a27f12f-scripts\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.495682 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sm7nq\" (UniqueName: \"kubernetes.io/projected/10ce3a20-645d-4c51-9736-87104a27f12f-kube-api-access-sm7nq\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.495719 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.495725 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.495739 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10ce3a20-645d-4c51-9736-87104a27f12f-config-data\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.495807 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-dev\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.495873 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-lib-modules\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.495884 4742 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") device mount path \"/mnt/openstack/pv04\"" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.495943 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-dev\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.495954 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-lib-modules\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.495973 4742 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") device mount path \"/mnt/openstack/pv02\"" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.496190 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/10ce3a20-645d-4c51-9736-87104a27f12f-httpd-run\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.496204 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-sys\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.496243 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-etc-nvme\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.496253 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10ce3a20-645d-4c51-9736-87104a27f12f-logs\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.496247 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-sys\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.496315 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-etc-nvme\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.496367 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.496403 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-run\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.496473 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.496525 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-run\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.505362 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10ce3a20-645d-4c51-9736-87104a27f12f-config-data\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.511239 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10ce3a20-645d-4c51-9736-87104a27f12f-scripts\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.519846 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sm7nq\" (UniqueName: \"kubernetes.io/projected/10ce3a20-645d-4c51-9736-87104a27f12f-kube-api-access-sm7nq\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.523255 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.542567 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-single-1\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.612909 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:27:57 crc kubenswrapper[4742]: I1004 03:27:57.906200 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 04 03:27:58 crc kubenswrapper[4742]: I1004 03:27:58.214513 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"10ce3a20-645d-4c51-9736-87104a27f12f","Type":"ContainerStarted","Data":"4da6e8311b78eada966738a47b27d003b7d4023f5bc8742c143b6302fd932c4b"} Oct 04 03:27:58 crc kubenswrapper[4742]: I1004 03:27:58.215393 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"10ce3a20-645d-4c51-9736-87104a27f12f","Type":"ContainerStarted","Data":"89dc862d64a860a7c9ffbbc159ecf5877d5364c8f39d17845c185ac1d0629aaf"} Oct 04 03:27:58 crc kubenswrapper[4742]: I1004 03:27:58.868375 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bfe21a9-2af5-43c3-b566-3d4529e4e542" path="/var/lib/kubelet/pods/3bfe21a9-2af5-43c3-b566-3d4529e4e542/volumes" Oct 04 03:27:59 crc kubenswrapper[4742]: I1004 03:27:59.231394 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"10ce3a20-645d-4c51-9736-87104a27f12f","Type":"ContainerStarted","Data":"74e1640ede2ac854978e46ed3b59d15f6f5edd84c262502f7b9b0c92e808ba44"} Oct 04 03:27:59 crc kubenswrapper[4742]: I1004 03:27:59.270821 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-1" podStartSLOduration=2.2707918 podStartE2EDuration="2.2707918s" podCreationTimestamp="2025-10-04 03:27:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:27:59.267539627 +0000 UTC m=+1045.199876843" watchObservedRunningTime="2025-10-04 03:27:59.2707918 +0000 UTC m=+1045.203129006" Oct 04 03:28:05 crc kubenswrapper[4742]: I1004 03:28:05.174323 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:05 crc kubenswrapper[4742]: I1004 03:28:05.175368 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:05 crc kubenswrapper[4742]: I1004 03:28:05.230714 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:05 crc kubenswrapper[4742]: I1004 03:28:05.249603 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:05 crc kubenswrapper[4742]: I1004 03:28:05.293686 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:05 crc kubenswrapper[4742]: I1004 03:28:05.294048 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:07 crc kubenswrapper[4742]: I1004 03:28:07.614151 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:28:07 crc kubenswrapper[4742]: I1004 03:28:07.614809 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:28:07 crc kubenswrapper[4742]: I1004 03:28:07.663978 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:28:07 crc kubenswrapper[4742]: I1004 03:28:07.683229 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:28:07 crc kubenswrapper[4742]: I1004 03:28:07.986911 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:07 crc kubenswrapper[4742]: I1004 03:28:07.987910 4742 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 03:28:08 crc kubenswrapper[4742]: I1004 03:28:08.090682 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:08 crc kubenswrapper[4742]: I1004 03:28:08.316842 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:28:08 crc kubenswrapper[4742]: I1004 03:28:08.316909 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:28:10 crc kubenswrapper[4742]: I1004 03:28:10.442745 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:28:10 crc kubenswrapper[4742]: I1004 03:28:10.443479 4742 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 03:28:10 crc kubenswrapper[4742]: I1004 03:28:10.518781 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:28:10 crc kubenswrapper[4742]: I1004 03:28:10.564041 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 03:28:10 crc kubenswrapper[4742]: I1004 03:28:10.564344 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="db88c4d6-966a-4144-90a8-33b468f422aa" containerName="glance-log" containerID="cri-o://cad70c7c3e4c0335aeb00f9b6533f96dbf1092bd9fe6818171b995731586a710" gracePeriod=30 Oct 04 03:28:10 crc kubenswrapper[4742]: I1004 03:28:10.564435 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="db88c4d6-966a-4144-90a8-33b468f422aa" containerName="glance-httpd" containerID="cri-o://4e7eafaddf997c85b637881517f7ff64e36fa44aff7a02665ed2cb33e59ba606" gracePeriod=30 Oct 04 03:28:11 crc kubenswrapper[4742]: I1004 03:28:11.351098 4742 generic.go:334] "Generic (PLEG): container finished" podID="db88c4d6-966a-4144-90a8-33b468f422aa" containerID="cad70c7c3e4c0335aeb00f9b6533f96dbf1092bd9fe6818171b995731586a710" exitCode=143 Oct 04 03:28:11 crc kubenswrapper[4742]: I1004 03:28:11.351245 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"db88c4d6-966a-4144-90a8-33b468f422aa","Type":"ContainerDied","Data":"cad70c7c3e4c0335aeb00f9b6533f96dbf1092bd9fe6818171b995731586a710"} Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.385015 4742 generic.go:334] "Generic (PLEG): container finished" podID="db88c4d6-966a-4144-90a8-33b468f422aa" containerID="4e7eafaddf997c85b637881517f7ff64e36fa44aff7a02665ed2cb33e59ba606" exitCode=0 Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.385124 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"db88c4d6-966a-4144-90a8-33b468f422aa","Type":"ContainerDied","Data":"4e7eafaddf997c85b637881517f7ff64e36fa44aff7a02665ed2cb33e59ba606"} Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.778563 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.871842 4742 patch_prober.go:28] interesting pod/machine-config-daemon-qdt7v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.872443 4742 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.872500 4742 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.873197 4742 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"85a4c342170ba17dd23c083b2ba6f556cfcde17e4bf847c7bc6921140b4883ad"} pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.873255 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerName="machine-config-daemon" containerID="cri-o://85a4c342170ba17dd23c083b2ba6f556cfcde17e4bf847c7bc6921140b4883ad" gracePeriod=600 Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.896791 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-etc-nvme\") pod \"db88c4d6-966a-4144-90a8-33b468f422aa\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.896856 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-dev\") pod \"db88c4d6-966a-4144-90a8-33b468f422aa\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.896912 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "db88c4d6-966a-4144-90a8-33b468f422aa" (UID: "db88c4d6-966a-4144-90a8-33b468f422aa"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.897013 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-dev" (OuterVolumeSpecName: "dev") pod "db88c4d6-966a-4144-90a8-33b468f422aa" (UID: "db88c4d6-966a-4144-90a8-33b468f422aa"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.896924 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-etc-iscsi\") pod \"db88c4d6-966a-4144-90a8-33b468f422aa\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.897065 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "db88c4d6-966a-4144-90a8-33b468f422aa" (UID: "db88c4d6-966a-4144-90a8-33b468f422aa"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.897096 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/db88c4d6-966a-4144-90a8-33b468f422aa-httpd-run\") pod \"db88c4d6-966a-4144-90a8-33b468f422aa\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.897490 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db88c4d6-966a-4144-90a8-33b468f422aa-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "db88c4d6-966a-4144-90a8-33b468f422aa" (UID: "db88c4d6-966a-4144-90a8-33b468f422aa"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.897690 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-lib-modules\") pod \"db88c4d6-966a-4144-90a8-33b468f422aa\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.897766 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db88c4d6-966a-4144-90a8-33b468f422aa-scripts\") pod \"db88c4d6-966a-4144-90a8-33b468f422aa\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.897810 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-var-locks-brick\") pod \"db88c4d6-966a-4144-90a8-33b468f422aa\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.897846 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"db88c4d6-966a-4144-90a8-33b468f422aa\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.897881 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-run\") pod \"db88c4d6-966a-4144-90a8-33b468f422aa\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.897814 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "db88c4d6-966a-4144-90a8-33b468f422aa" (UID: "db88c4d6-966a-4144-90a8-33b468f422aa"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.897898 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "db88c4d6-966a-4144-90a8-33b468f422aa" (UID: "db88c4d6-966a-4144-90a8-33b468f422aa"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.897918 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db88c4d6-966a-4144-90a8-33b468f422aa-config-data\") pod \"db88c4d6-966a-4144-90a8-33b468f422aa\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.897941 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-sys\") pod \"db88c4d6-966a-4144-90a8-33b468f422aa\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.897953 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-run" (OuterVolumeSpecName: "run") pod "db88c4d6-966a-4144-90a8-33b468f422aa" (UID: "db88c4d6-966a-4144-90a8-33b468f422aa"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.897991 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jjbsv\" (UniqueName: \"kubernetes.io/projected/db88c4d6-966a-4144-90a8-33b468f422aa-kube-api-access-jjbsv\") pod \"db88c4d6-966a-4144-90a8-33b468f422aa\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.898016 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db88c4d6-966a-4144-90a8-33b468f422aa-logs\") pod \"db88c4d6-966a-4144-90a8-33b468f422aa\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.898043 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"db88c4d6-966a-4144-90a8-33b468f422aa\" (UID: \"db88c4d6-966a-4144-90a8-33b468f422aa\") " Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.898104 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-sys" (OuterVolumeSpecName: "sys") pod "db88c4d6-966a-4144-90a8-33b468f422aa" (UID: "db88c4d6-966a-4144-90a8-33b468f422aa"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.898933 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db88c4d6-966a-4144-90a8-33b468f422aa-logs" (OuterVolumeSpecName: "logs") pod "db88c4d6-966a-4144-90a8-33b468f422aa" (UID: "db88c4d6-966a-4144-90a8-33b468f422aa"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.899869 4742 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-etc-nvme\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.899908 4742 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-dev\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.899930 4742 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-etc-iscsi\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.899948 4742 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/db88c4d6-966a-4144-90a8-33b468f422aa-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.899965 4742 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-lib-modules\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.899984 4742 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-var-locks-brick\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.900003 4742 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-run\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.900023 4742 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/db88c4d6-966a-4144-90a8-33b468f422aa-sys\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.900040 4742 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db88c4d6-966a-4144-90a8-33b468f422aa-logs\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.904045 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "db88c4d6-966a-4144-90a8-33b468f422aa" (UID: "db88c4d6-966a-4144-90a8-33b468f422aa"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.904743 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance-cache") pod "db88c4d6-966a-4144-90a8-33b468f422aa" (UID: "db88c4d6-966a-4144-90a8-33b468f422aa"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.914214 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db88c4d6-966a-4144-90a8-33b468f422aa-scripts" (OuterVolumeSpecName: "scripts") pod "db88c4d6-966a-4144-90a8-33b468f422aa" (UID: "db88c4d6-966a-4144-90a8-33b468f422aa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.928075 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db88c4d6-966a-4144-90a8-33b468f422aa-kube-api-access-jjbsv" (OuterVolumeSpecName: "kube-api-access-jjbsv") pod "db88c4d6-966a-4144-90a8-33b468f422aa" (UID: "db88c4d6-966a-4144-90a8-33b468f422aa"). InnerVolumeSpecName "kube-api-access-jjbsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:28:14 crc kubenswrapper[4742]: I1004 03:28:14.947871 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db88c4d6-966a-4144-90a8-33b468f422aa-config-data" (OuterVolumeSpecName: "config-data") pod "db88c4d6-966a-4144-90a8-33b468f422aa" (UID: "db88c4d6-966a-4144-90a8-33b468f422aa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.003924 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jjbsv\" (UniqueName: \"kubernetes.io/projected/db88c4d6-966a-4144-90a8-33b468f422aa-kube-api-access-jjbsv\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.004104 4742 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.004183 4742 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db88c4d6-966a-4144-90a8-33b468f422aa-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.004254 4742 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.004400 4742 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db88c4d6-966a-4144-90a8-33b468f422aa-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.018298 4742 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.022221 4742 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.106292 4742 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.106322 4742 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.397008 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.397060 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"db88c4d6-966a-4144-90a8-33b468f422aa","Type":"ContainerDied","Data":"f509931cffc450cb914a27a6798487fd6de77d4635949c66637149ed385aaefb"} Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.397695 4742 scope.go:117] "RemoveContainer" containerID="4e7eafaddf997c85b637881517f7ff64e36fa44aff7a02665ed2cb33e59ba606" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.404314 4742 generic.go:334] "Generic (PLEG): container finished" podID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerID="85a4c342170ba17dd23c083b2ba6f556cfcde17e4bf847c7bc6921140b4883ad" exitCode=0 Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.404353 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" event={"ID":"989ea247-7ac9-4ade-8dcb-2632fef4f83e","Type":"ContainerDied","Data":"85a4c342170ba17dd23c083b2ba6f556cfcde17e4bf847c7bc6921140b4883ad"} Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.404380 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" event={"ID":"989ea247-7ac9-4ade-8dcb-2632fef4f83e","Type":"ContainerStarted","Data":"97c026eaa827ea7984bf1c754436a808d45b23a630beb167b6e8d3d8d75609e3"} Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.434156 4742 scope.go:117] "RemoveContainer" containerID="cad70c7c3e4c0335aeb00f9b6533f96dbf1092bd9fe6818171b995731586a710" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.446990 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.451596 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.458524 4742 scope.go:117] "RemoveContainer" containerID="ffcec0a1f833323e5678d3dde9b3cc1a6ca05ee38cfe04eae68bdf431a24a5f2" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.480849 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 03:28:15 crc kubenswrapper[4742]: E1004 03:28:15.481349 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db88c4d6-966a-4144-90a8-33b468f422aa" containerName="glance-log" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.481372 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="db88c4d6-966a-4144-90a8-33b468f422aa" containerName="glance-log" Oct 04 03:28:15 crc kubenswrapper[4742]: E1004 03:28:15.481393 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db88c4d6-966a-4144-90a8-33b468f422aa" containerName="glance-httpd" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.481400 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="db88c4d6-966a-4144-90a8-33b468f422aa" containerName="glance-httpd" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.481552 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="db88c4d6-966a-4144-90a8-33b468f422aa" containerName="glance-log" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.481588 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="db88c4d6-966a-4144-90a8-33b468f422aa" containerName="glance-httpd" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.484262 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.500120 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.613148 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.613468 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-httpd-run\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.613579 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-dev\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.613695 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rd955\" (UniqueName: \"kubernetes.io/projected/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-kube-api-access-rd955\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.613770 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-logs\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.613855 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-lib-modules\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.613930 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-sys\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.614012 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.614092 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-run\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.614194 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-etc-nvme\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.614292 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.614397 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-config-data\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.614473 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-scripts\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.614567 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.715931 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.716401 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.716434 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-httpd-run\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.716466 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-dev\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.716504 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rd955\" (UniqueName: \"kubernetes.io/projected/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-kube-api-access-rd955\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.716523 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-logs\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.716545 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-lib-modules\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.716564 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-sys\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.716583 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.716605 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-run\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.716647 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-etc-nvme\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.716674 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.716702 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-config-data\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.716724 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-scripts\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.716746 4742 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") device mount path \"/mnt/openstack/pv05\"" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.716994 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-lib-modules\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.717159 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-sys\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.717231 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.717260 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-run\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.717442 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-etc-nvme\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.717549 4742 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.717612 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-httpd-run\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.716122 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.717677 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-dev\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.718414 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-logs\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.733029 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-scripts\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.733182 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-config-data\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.737430 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.739559 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.742381 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rd955\" (UniqueName: \"kubernetes.io/projected/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-kube-api-access-rd955\") pod \"glance-default-single-0\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:15 crc kubenswrapper[4742]: I1004 03:28:15.852512 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:16 crc kubenswrapper[4742]: I1004 03:28:16.120430 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 03:28:16 crc kubenswrapper[4742]: I1004 03:28:16.420887 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4","Type":"ContainerStarted","Data":"9bde0eeee13129290da855978b66c295dc0998a39050d3c4f7bc3a992b1925d9"} Oct 04 03:28:16 crc kubenswrapper[4742]: I1004 03:28:16.421389 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4","Type":"ContainerStarted","Data":"df08d74e355c0363dd010aa1b8a8385b4feed55fca91cd6cb8d020197d59915a"} Oct 04 03:28:16 crc kubenswrapper[4742]: I1004 03:28:16.872644 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db88c4d6-966a-4144-90a8-33b468f422aa" path="/var/lib/kubelet/pods/db88c4d6-966a-4144-90a8-33b468f422aa/volumes" Oct 04 03:28:17 crc kubenswrapper[4742]: I1004 03:28:17.433306 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4","Type":"ContainerStarted","Data":"583d124836802f60be61e94bf76c355ed921bd0847d9db19f8bcf65a32922c71"} Oct 04 03:28:17 crc kubenswrapper[4742]: I1004 03:28:17.471386 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=2.47135928 podStartE2EDuration="2.47135928s" podCreationTimestamp="2025-10-04 03:28:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:28:17.465505719 +0000 UTC m=+1063.397842925" watchObservedRunningTime="2025-10-04 03:28:17.47135928 +0000 UTC m=+1063.403696496" Oct 04 03:28:25 crc kubenswrapper[4742]: I1004 03:28:25.853403 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:25 crc kubenswrapper[4742]: I1004 03:28:25.854654 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:25 crc kubenswrapper[4742]: I1004 03:28:25.888730 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:25 crc kubenswrapper[4742]: I1004 03:28:25.899302 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:26 crc kubenswrapper[4742]: I1004 03:28:26.523235 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:26 crc kubenswrapper[4742]: I1004 03:28:26.523467 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:28 crc kubenswrapper[4742]: I1004 03:28:28.449875 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:28 crc kubenswrapper[4742]: I1004 03:28:28.453877 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.121633 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-2dq5q"] Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.130646 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-2dq5q"] Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.167494 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance52a1-account-delete-p44vn"] Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.168987 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance52a1-account-delete-p44vn" Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.185818 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.186123 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="db5f30ca-ef98-4f4e-97e6-d3880d6f31a4" containerName="glance-log" containerID="cri-o://9bde0eeee13129290da855978b66c295dc0998a39050d3c4f7bc3a992b1925d9" gracePeriod=30 Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.186325 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="db5f30ca-ef98-4f4e-97e6-d3880d6f31a4" containerName="glance-httpd" containerID="cri-o://583d124836802f60be61e94bf76c355ed921bd0847d9db19f8bcf65a32922c71" gracePeriod=30 Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.209548 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance52a1-account-delete-p44vn"] Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.209604 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.209795 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="10ce3a20-645d-4c51-9736-87104a27f12f" containerName="glance-log" containerID="cri-o://4da6e8311b78eada966738a47b27d003b7d4023f5bc8742c143b6302fd932c4b" gracePeriod=30 Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.209938 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="10ce3a20-645d-4c51-9736-87104a27f12f" containerName="glance-httpd" containerID="cri-o://74e1640ede2ac854978e46ed3b59d15f6f5edd84c262502f7b9b0c92e808ba44" gracePeriod=30 Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.233155 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k4b6\" (UniqueName: \"kubernetes.io/projected/ab7fff03-3721-423c-9a11-8db657c1bde6-kube-api-access-9k4b6\") pod \"glance52a1-account-delete-p44vn\" (UID: \"ab7fff03-3721-423c-9a11-8db657c1bde6\") " pod="glance-kuttl-tests/glance52a1-account-delete-p44vn" Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.276395 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstackclient"] Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.277985 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/openstackclient" podUID="8d7e48d9-5aba-44f1-a31f-fb353d8e034d" containerName="openstackclient" containerID="cri-o://2c5d6c5186fcc36d6be9f9b3c45e55eb3c383e9f0d1bae92257f78ab051b135f" gracePeriod=30 Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.334768 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9k4b6\" (UniqueName: \"kubernetes.io/projected/ab7fff03-3721-423c-9a11-8db657c1bde6-kube-api-access-9k4b6\") pod \"glance52a1-account-delete-p44vn\" (UID: \"ab7fff03-3721-423c-9a11-8db657c1bde6\") " pod="glance-kuttl-tests/glance52a1-account-delete-p44vn" Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.357773 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9k4b6\" (UniqueName: \"kubernetes.io/projected/ab7fff03-3721-423c-9a11-8db657c1bde6-kube-api-access-9k4b6\") pod \"glance52a1-account-delete-p44vn\" (UID: \"ab7fff03-3721-423c-9a11-8db657c1bde6\") " pod="glance-kuttl-tests/glance52a1-account-delete-p44vn" Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.548995 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance52a1-account-delete-p44vn" Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.687763 4742 generic.go:334] "Generic (PLEG): container finished" podID="10ce3a20-645d-4c51-9736-87104a27f12f" containerID="4da6e8311b78eada966738a47b27d003b7d4023f5bc8742c143b6302fd932c4b" exitCode=143 Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.688300 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"10ce3a20-645d-4c51-9736-87104a27f12f","Type":"ContainerDied","Data":"4da6e8311b78eada966738a47b27d003b7d4023f5bc8742c143b6302fd932c4b"} Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.697883 4742 generic.go:334] "Generic (PLEG): container finished" podID="db5f30ca-ef98-4f4e-97e6-d3880d6f31a4" containerID="9bde0eeee13129290da855978b66c295dc0998a39050d3c4f7bc3a992b1925d9" exitCode=143 Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.697932 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4","Type":"ContainerDied","Data":"9bde0eeee13129290da855978b66c295dc0998a39050d3c4f7bc3a992b1925d9"} Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.704538 4742 generic.go:334] "Generic (PLEG): container finished" podID="8d7e48d9-5aba-44f1-a31f-fb353d8e034d" containerID="2c5d6c5186fcc36d6be9f9b3c45e55eb3c383e9f0d1bae92257f78ab051b135f" exitCode=143 Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.704593 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"8d7e48d9-5aba-44f1-a31f-fb353d8e034d","Type":"ContainerDied","Data":"2c5d6c5186fcc36d6be9f9b3c45e55eb3c383e9f0d1bae92257f78ab051b135f"} Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.704614 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"8d7e48d9-5aba-44f1-a31f-fb353d8e034d","Type":"ContainerDied","Data":"720e3c018441a8e99cebcf93f5cbcace6a3cd9433e8c20fb4afbfef51970794f"} Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.704626 4742 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="720e3c018441a8e99cebcf93f5cbcace6a3cd9433e8c20fb4afbfef51970794f" Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.712605 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.826852 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance52a1-account-delete-p44vn"] Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.842670 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-47zlz\" (UniqueName: \"kubernetes.io/projected/8d7e48d9-5aba-44f1-a31f-fb353d8e034d-kube-api-access-47zlz\") pod \"8d7e48d9-5aba-44f1-a31f-fb353d8e034d\" (UID: \"8d7e48d9-5aba-44f1-a31f-fb353d8e034d\") " Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.842751 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8d7e48d9-5aba-44f1-a31f-fb353d8e034d-openstack-config-secret\") pod \"8d7e48d9-5aba-44f1-a31f-fb353d8e034d\" (UID: \"8d7e48d9-5aba-44f1-a31f-fb353d8e034d\") " Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.842953 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8d7e48d9-5aba-44f1-a31f-fb353d8e034d-openstack-config\") pod \"8d7e48d9-5aba-44f1-a31f-fb353d8e034d\" (UID: \"8d7e48d9-5aba-44f1-a31f-fb353d8e034d\") " Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.842983 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/8d7e48d9-5aba-44f1-a31f-fb353d8e034d-openstack-scripts\") pod \"8d7e48d9-5aba-44f1-a31f-fb353d8e034d\" (UID: \"8d7e48d9-5aba-44f1-a31f-fb353d8e034d\") " Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.844128 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d7e48d9-5aba-44f1-a31f-fb353d8e034d-openstack-scripts" (OuterVolumeSpecName: "openstack-scripts") pod "8d7e48d9-5aba-44f1-a31f-fb353d8e034d" (UID: "8d7e48d9-5aba-44f1-a31f-fb353d8e034d"). InnerVolumeSpecName "openstack-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.848776 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d7e48d9-5aba-44f1-a31f-fb353d8e034d-kube-api-access-47zlz" (OuterVolumeSpecName: "kube-api-access-47zlz") pod "8d7e48d9-5aba-44f1-a31f-fb353d8e034d" (UID: "8d7e48d9-5aba-44f1-a31f-fb353d8e034d"). InnerVolumeSpecName "kube-api-access-47zlz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.866452 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d7e48d9-5aba-44f1-a31f-fb353d8e034d-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "8d7e48d9-5aba-44f1-a31f-fb353d8e034d" (UID: "8d7e48d9-5aba-44f1-a31f-fb353d8e034d"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.871150 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d7e48d9-5aba-44f1-a31f-fb353d8e034d-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "8d7e48d9-5aba-44f1-a31f-fb353d8e034d" (UID: "8d7e48d9-5aba-44f1-a31f-fb353d8e034d"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.945718 4742 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8d7e48d9-5aba-44f1-a31f-fb353d8e034d-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.945785 4742 reconciler_common.go:293] "Volume detached for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/8d7e48d9-5aba-44f1-a31f-fb353d8e034d-openstack-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.945801 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-47zlz\" (UniqueName: \"kubernetes.io/projected/8d7e48d9-5aba-44f1-a31f-fb353d8e034d-kube-api-access-47zlz\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:43 crc kubenswrapper[4742]: I1004 03:28:43.945815 4742 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8d7e48d9-5aba-44f1-a31f-fb353d8e034d-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:44 crc kubenswrapper[4742]: I1004 03:28:44.716957 4742 generic.go:334] "Generic (PLEG): container finished" podID="ab7fff03-3721-423c-9a11-8db657c1bde6" containerID="a1df9f270da043479282e7b102d2d7dc737edc597d3d0308489d98f5b11df8cf" exitCode=0 Oct 04 03:28:44 crc kubenswrapper[4742]: I1004 03:28:44.717006 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance52a1-account-delete-p44vn" event={"ID":"ab7fff03-3721-423c-9a11-8db657c1bde6","Type":"ContainerDied","Data":"a1df9f270da043479282e7b102d2d7dc737edc597d3d0308489d98f5b11df8cf"} Oct 04 03:28:44 crc kubenswrapper[4742]: I1004 03:28:44.718422 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance52a1-account-delete-p44vn" event={"ID":"ab7fff03-3721-423c-9a11-8db657c1bde6","Type":"ContainerStarted","Data":"8ef96aa167b9e621234e636f1df7c422f110cf9a48743b2fc9c57660e20d5cb8"} Oct 04 03:28:44 crc kubenswrapper[4742]: I1004 03:28:44.718570 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Oct 04 03:28:44 crc kubenswrapper[4742]: I1004 03:28:44.773062 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstackclient"] Oct 04 03:28:44 crc kubenswrapper[4742]: I1004 03:28:44.789557 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/openstackclient"] Oct 04 03:28:44 crc kubenswrapper[4742]: I1004 03:28:44.866874 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d7e48d9-5aba-44f1-a31f-fb353d8e034d" path="/var/lib/kubelet/pods/8d7e48d9-5aba-44f1-a31f-fb353d8e034d/volumes" Oct 04 03:28:44 crc kubenswrapper[4742]: I1004 03:28:44.867616 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6" path="/var/lib/kubelet/pods/b583f1b2-1e25-4f72-abf3-c79a7eaf6cb6/volumes" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.145394 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance52a1-account-delete-p44vn" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.186262 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9k4b6\" (UniqueName: \"kubernetes.io/projected/ab7fff03-3721-423c-9a11-8db657c1bde6-kube-api-access-9k4b6\") pod \"ab7fff03-3721-423c-9a11-8db657c1bde6\" (UID: \"ab7fff03-3721-423c-9a11-8db657c1bde6\") " Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.194915 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab7fff03-3721-423c-9a11-8db657c1bde6-kube-api-access-9k4b6" (OuterVolumeSpecName: "kube-api-access-9k4b6") pod "ab7fff03-3721-423c-9a11-8db657c1bde6" (UID: "ab7fff03-3721-423c-9a11-8db657c1bde6"). InnerVolumeSpecName "kube-api-access-9k4b6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.288577 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9k4b6\" (UniqueName: \"kubernetes.io/projected/ab7fff03-3721-423c-9a11-8db657c1bde6-kube-api-access-9k4b6\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.407901 4742 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="db5f30ca-ef98-4f4e-97e6-d3880d6f31a4" containerName="glance-log" probeResult="failure" output="Get \"http://10.217.0.105:9292/healthcheck\": read tcp 10.217.0.2:38284->10.217.0.105:9292: read: connection reset by peer" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.407983 4742 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="db5f30ca-ef98-4f4e-97e6-d3880d6f31a4" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.0.105:9292/healthcheck\": read tcp 10.217.0.2:38282->10.217.0.105:9292: read: connection reset by peer" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.757213 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance52a1-account-delete-p44vn" event={"ID":"ab7fff03-3721-423c-9a11-8db657c1bde6","Type":"ContainerDied","Data":"8ef96aa167b9e621234e636f1df7c422f110cf9a48743b2fc9c57660e20d5cb8"} Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.757768 4742 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ef96aa167b9e621234e636f1df7c422f110cf9a48743b2fc9c57660e20d5cb8" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.757370 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance52a1-account-delete-p44vn" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.759855 4742 generic.go:334] "Generic (PLEG): container finished" podID="10ce3a20-645d-4c51-9736-87104a27f12f" containerID="74e1640ede2ac854978e46ed3b59d15f6f5edd84c262502f7b9b0c92e808ba44" exitCode=0 Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.759950 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"10ce3a20-645d-4c51-9736-87104a27f12f","Type":"ContainerDied","Data":"74e1640ede2ac854978e46ed3b59d15f6f5edd84c262502f7b9b0c92e808ba44"} Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.763008 4742 generic.go:334] "Generic (PLEG): container finished" podID="db5f30ca-ef98-4f4e-97e6-d3880d6f31a4" containerID="583d124836802f60be61e94bf76c355ed921bd0847d9db19f8bcf65a32922c71" exitCode=0 Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.763066 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4","Type":"ContainerDied","Data":"583d124836802f60be61e94bf76c355ed921bd0847d9db19f8bcf65a32922c71"} Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.771164 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.833143 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.897369 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-run\") pod \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.898854 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-lib-modules\") pod \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.898973 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10ce3a20-645d-4c51-9736-87104a27f12f-config-data\") pod \"10ce3a20-645d-4c51-9736-87104a27f12f\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.899102 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10ce3a20-645d-4c51-9736-87104a27f12f-scripts\") pod \"10ce3a20-645d-4c51-9736-87104a27f12f\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.899212 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rd955\" (UniqueName: \"kubernetes.io/projected/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-kube-api-access-rd955\") pod \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.899484 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-logs\") pod \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.899666 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-etc-iscsi\") pod \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.899911 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"10ce3a20-645d-4c51-9736-87104a27f12f\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.900027 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-config-data\") pod \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.900114 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-var-locks-brick\") pod \"10ce3a20-645d-4c51-9736-87104a27f12f\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.900230 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/10ce3a20-645d-4c51-9736-87104a27f12f-httpd-run\") pod \"10ce3a20-645d-4c51-9736-87104a27f12f\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.900371 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-lib-modules\") pod \"10ce3a20-645d-4c51-9736-87104a27f12f\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.900466 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.900539 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-run\") pod \"10ce3a20-645d-4c51-9736-87104a27f12f\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.900663 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-dev\") pod \"10ce3a20-645d-4c51-9736-87104a27f12f\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.901460 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sm7nq\" (UniqueName: \"kubernetes.io/projected/10ce3a20-645d-4c51-9736-87104a27f12f-kube-api-access-sm7nq\") pod \"10ce3a20-645d-4c51-9736-87104a27f12f\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.901556 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-httpd-run\") pod \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.901625 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-var-locks-brick\") pod \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.902716 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-scripts\") pod \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.902911 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"10ce3a20-645d-4c51-9736-87104a27f12f\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.898129 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-run" (OuterVolumeSpecName: "run") pod "db5f30ca-ef98-4f4e-97e6-d3880d6f31a4" (UID: "db5f30ca-ef98-4f4e-97e6-d3880d6f31a4"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.898955 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "db5f30ca-ef98-4f4e-97e6-d3880d6f31a4" (UID: "db5f30ca-ef98-4f4e-97e6-d3880d6f31a4"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.899758 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "db5f30ca-ef98-4f4e-97e6-d3880d6f31a4" (UID: "db5f30ca-ef98-4f4e-97e6-d3880d6f31a4"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.899962 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-logs" (OuterVolumeSpecName: "logs") pod "db5f30ca-ef98-4f4e-97e6-d3880d6f31a4" (UID: "db5f30ca-ef98-4f4e-97e6-d3880d6f31a4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.904115 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "10ce3a20-645d-4c51-9736-87104a27f12f" (UID: "10ce3a20-645d-4c51-9736-87104a27f12f"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.904165 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-run" (OuterVolumeSpecName: "run") pod "10ce3a20-645d-4c51-9736-87104a27f12f" (UID: "10ce3a20-645d-4c51-9736-87104a27f12f"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.904190 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "10ce3a20-645d-4c51-9736-87104a27f12f" (UID: "10ce3a20-645d-4c51-9736-87104a27f12f"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.904463 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10ce3a20-645d-4c51-9736-87104a27f12f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "10ce3a20-645d-4c51-9736-87104a27f12f" (UID: "10ce3a20-645d-4c51-9736-87104a27f12f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.904497 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "10ce3a20-645d-4c51-9736-87104a27f12f" (UID: "10ce3a20-645d-4c51-9736-87104a27f12f"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.905386 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-kube-api-access-rd955" (OuterVolumeSpecName: "kube-api-access-rd955") pod "db5f30ca-ef98-4f4e-97e6-d3880d6f31a4" (UID: "db5f30ca-ef98-4f4e-97e6-d3880d6f31a4"). InnerVolumeSpecName "kube-api-access-rd955". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.905702 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "db5f30ca-ef98-4f4e-97e6-d3880d6f31a4" (UID: "db5f30ca-ef98-4f4e-97e6-d3880d6f31a4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.905734 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-dev" (OuterVolumeSpecName: "dev") pod "10ce3a20-645d-4c51-9736-87104a27f12f" (UID: "10ce3a20-645d-4c51-9736-87104a27f12f"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.905921 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10ce3a20-645d-4c51-9736-87104a27f12f-scripts" (OuterVolumeSpecName: "scripts") pod "10ce3a20-645d-4c51-9736-87104a27f12f" (UID: "10ce3a20-645d-4c51-9736-87104a27f12f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.905967 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "db5f30ca-ef98-4f4e-97e6-d3880d6f31a4" (UID: "db5f30ca-ef98-4f4e-97e6-d3880d6f31a4"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.907533 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance-cache") pod "db5f30ca-ef98-4f4e-97e6-d3880d6f31a4" (UID: "db5f30ca-ef98-4f4e-97e6-d3880d6f31a4"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.909631 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10ce3a20-645d-4c51-9736-87104a27f12f-kube-api-access-sm7nq" (OuterVolumeSpecName: "kube-api-access-sm7nq") pod "10ce3a20-645d-4c51-9736-87104a27f12f" (UID: "10ce3a20-645d-4c51-9736-87104a27f12f"). InnerVolumeSpecName "kube-api-access-sm7nq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.910590 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-scripts" (OuterVolumeSpecName: "scripts") pod "db5f30ca-ef98-4f4e-97e6-d3880d6f31a4" (UID: "db5f30ca-ef98-4f4e-97e6-d3880d6f31a4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.911638 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance-cache") pod "10ce3a20-645d-4c51-9736-87104a27f12f" (UID: "10ce3a20-645d-4c51-9736-87104a27f12f"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.913610 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "db5f30ca-ef98-4f4e-97e6-d3880d6f31a4" (UID: "db5f30ca-ef98-4f4e-97e6-d3880d6f31a4"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.913544 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-etc-nvme\") pod \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.914327 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.915590 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-sys\") pod \"10ce3a20-645d-4c51-9736-87104a27f12f\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.915752 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-etc-nvme\") pod \"10ce3a20-645d-4c51-9736-87104a27f12f\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.915907 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10ce3a20-645d-4c51-9736-87104a27f12f-logs\") pod \"10ce3a20-645d-4c51-9736-87104a27f12f\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.916585 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-dev\") pod \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.915649 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-sys" (OuterVolumeSpecName: "sys") pod "10ce3a20-645d-4c51-9736-87104a27f12f" (UID: "10ce3a20-645d-4c51-9736-87104a27f12f"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.915820 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "10ce3a20-645d-4c51-9736-87104a27f12f" (UID: "10ce3a20-645d-4c51-9736-87104a27f12f"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.916344 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10ce3a20-645d-4c51-9736-87104a27f12f-logs" (OuterVolumeSpecName: "logs") pod "10ce3a20-645d-4c51-9736-87104a27f12f" (UID: "10ce3a20-645d-4c51-9736-87104a27f12f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.916659 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-dev" (OuterVolumeSpecName: "dev") pod "db5f30ca-ef98-4f4e-97e6-d3880d6f31a4" (UID: "db5f30ca-ef98-4f4e-97e6-d3880d6f31a4"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.916869 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "10ce3a20-645d-4c51-9736-87104a27f12f" (UID: "10ce3a20-645d-4c51-9736-87104a27f12f"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.916982 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-etc-iscsi\") pod \"10ce3a20-645d-4c51-9736-87104a27f12f\" (UID: \"10ce3a20-645d-4c51-9736-87104a27f12f\") " Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.917079 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-sys\") pod \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\" (UID: \"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4\") " Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.917153 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-sys" (OuterVolumeSpecName: "sys") pod "db5f30ca-ef98-4f4e-97e6-d3880d6f31a4" (UID: "db5f30ca-ef98-4f4e-97e6-d3880d6f31a4"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.918126 4742 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-sys\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.918288 4742 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-etc-nvme\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.918380 4742 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10ce3a20-645d-4c51-9736-87104a27f12f-logs\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.918457 4742 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-dev\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.918536 4742 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-etc-iscsi\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.918613 4742 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-sys\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.918704 4742 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-run\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.918782 4742 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-lib-modules\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.918868 4742 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10ce3a20-645d-4c51-9736-87104a27f12f-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.918959 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rd955\" (UniqueName: \"kubernetes.io/projected/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-kube-api-access-rd955\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.919043 4742 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-logs\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.919114 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "db5f30ca-ef98-4f4e-97e6-d3880d6f31a4" (UID: "db5f30ca-ef98-4f4e-97e6-d3880d6f31a4"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.919122 4742 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-etc-iscsi\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.919225 4742 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.919302 4742 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-var-locks-brick\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.919330 4742 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/10ce3a20-645d-4c51-9736-87104a27f12f-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.919352 4742 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-lib-modules\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.919387 4742 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.919408 4742 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-run\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.919452 4742 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/10ce3a20-645d-4c51-9736-87104a27f12f-dev\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.919475 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sm7nq\" (UniqueName: \"kubernetes.io/projected/10ce3a20-645d-4c51-9736-87104a27f12f-kube-api-access-sm7nq\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.919496 4742 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.919517 4742 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.919538 4742 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-var-locks-brick\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.919568 4742 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.919596 4742 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-etc-nvme\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.934958 4742 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.938215 4742 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.942939 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10ce3a20-645d-4c51-9736-87104a27f12f-config-data" (OuterVolumeSpecName: "config-data") pod "10ce3a20-645d-4c51-9736-87104a27f12f" (UID: "10ce3a20-645d-4c51-9736-87104a27f12f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.946732 4742 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 04 03:28:46 crc kubenswrapper[4742]: I1004 03:28:46.950646 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-config-data" (OuterVolumeSpecName: "config-data") pod "db5f30ca-ef98-4f4e-97e6-d3880d6f31a4" (UID: "db5f30ca-ef98-4f4e-97e6-d3880d6f31a4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:28:47 crc kubenswrapper[4742]: I1004 03:28:47.023647 4742 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10ce3a20-645d-4c51-9736-87104a27f12f-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:47 crc kubenswrapper[4742]: I1004 03:28:47.023703 4742 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:47 crc kubenswrapper[4742]: I1004 03:28:47.023727 4742 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:47 crc kubenswrapper[4742]: I1004 03:28:47.023747 4742 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:47 crc kubenswrapper[4742]: I1004 03:28:47.023767 4742 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:47 crc kubenswrapper[4742]: I1004 03:28:47.023831 4742 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 04 03:28:47 crc kubenswrapper[4742]: I1004 03:28:47.050987 4742 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 04 03:28:47 crc kubenswrapper[4742]: I1004 03:28:47.127744 4742 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:47 crc kubenswrapper[4742]: I1004 03:28:47.778387 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"10ce3a20-645d-4c51-9736-87104a27f12f","Type":"ContainerDied","Data":"89dc862d64a860a7c9ffbbc159ecf5877d5364c8f39d17845c185ac1d0629aaf"} Oct 04 03:28:47 crc kubenswrapper[4742]: I1004 03:28:47.778463 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 03:28:47 crc kubenswrapper[4742]: I1004 03:28:47.778505 4742 scope.go:117] "RemoveContainer" containerID="74e1640ede2ac854978e46ed3b59d15f6f5edd84c262502f7b9b0c92e808ba44" Oct 04 03:28:47 crc kubenswrapper[4742]: I1004 03:28:47.782420 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"db5f30ca-ef98-4f4e-97e6-d3880d6f31a4","Type":"ContainerDied","Data":"df08d74e355c0363dd010aa1b8a8385b4feed55fca91cd6cb8d020197d59915a"} Oct 04 03:28:47 crc kubenswrapper[4742]: I1004 03:28:47.782526 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:28:47 crc kubenswrapper[4742]: I1004 03:28:47.823432 4742 scope.go:117] "RemoveContainer" containerID="4da6e8311b78eada966738a47b27d003b7d4023f5bc8742c143b6302fd932c4b" Oct 04 03:28:47 crc kubenswrapper[4742]: I1004 03:28:47.841510 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 04 03:28:47 crc kubenswrapper[4742]: I1004 03:28:47.851770 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 04 03:28:47 crc kubenswrapper[4742]: I1004 03:28:47.854377 4742 scope.go:117] "RemoveContainer" containerID="583d124836802f60be61e94bf76c355ed921bd0847d9db19f8bcf65a32922c71" Oct 04 03:28:47 crc kubenswrapper[4742]: I1004 03:28:47.866592 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 03:28:47 crc kubenswrapper[4742]: I1004 03:28:47.873673 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 03:28:47 crc kubenswrapper[4742]: I1004 03:28:47.881948 4742 scope.go:117] "RemoveContainer" containerID="9bde0eeee13129290da855978b66c295dc0998a39050d3c4f7bc3a992b1925d9" Oct 04 03:28:48 crc kubenswrapper[4742]: I1004 03:28:48.177882 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-mgmlq"] Oct 04 03:28:48 crc kubenswrapper[4742]: I1004 03:28:48.184390 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-mgmlq"] Oct 04 03:28:48 crc kubenswrapper[4742]: I1004 03:28:48.193655 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance52a1-account-delete-p44vn"] Oct 04 03:28:48 crc kubenswrapper[4742]: I1004 03:28:48.202250 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance52a1-account-delete-p44vn"] Oct 04 03:28:48 crc kubenswrapper[4742]: I1004 03:28:48.208944 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-52a1-account-create-8zdhh"] Oct 04 03:28:48 crc kubenswrapper[4742]: I1004 03:28:48.213343 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-52a1-account-create-8zdhh"] Oct 04 03:28:48 crc kubenswrapper[4742]: I1004 03:28:48.578873 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-96mj9"] Oct 04 03:28:48 crc kubenswrapper[4742]: E1004 03:28:48.579142 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10ce3a20-645d-4c51-9736-87104a27f12f" containerName="glance-log" Oct 04 03:28:48 crc kubenswrapper[4742]: I1004 03:28:48.579156 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="10ce3a20-645d-4c51-9736-87104a27f12f" containerName="glance-log" Oct 04 03:28:48 crc kubenswrapper[4742]: E1004 03:28:48.579166 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d7e48d9-5aba-44f1-a31f-fb353d8e034d" containerName="openstackclient" Oct 04 03:28:48 crc kubenswrapper[4742]: I1004 03:28:48.579171 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d7e48d9-5aba-44f1-a31f-fb353d8e034d" containerName="openstackclient" Oct 04 03:28:48 crc kubenswrapper[4742]: E1004 03:28:48.579186 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab7fff03-3721-423c-9a11-8db657c1bde6" containerName="mariadb-account-delete" Oct 04 03:28:48 crc kubenswrapper[4742]: I1004 03:28:48.579192 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab7fff03-3721-423c-9a11-8db657c1bde6" containerName="mariadb-account-delete" Oct 04 03:28:48 crc kubenswrapper[4742]: E1004 03:28:48.579204 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db5f30ca-ef98-4f4e-97e6-d3880d6f31a4" containerName="glance-httpd" Oct 04 03:28:48 crc kubenswrapper[4742]: I1004 03:28:48.579210 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="db5f30ca-ef98-4f4e-97e6-d3880d6f31a4" containerName="glance-httpd" Oct 04 03:28:48 crc kubenswrapper[4742]: E1004 03:28:48.579229 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10ce3a20-645d-4c51-9736-87104a27f12f" containerName="glance-httpd" Oct 04 03:28:48 crc kubenswrapper[4742]: I1004 03:28:48.579236 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="10ce3a20-645d-4c51-9736-87104a27f12f" containerName="glance-httpd" Oct 04 03:28:48 crc kubenswrapper[4742]: E1004 03:28:48.579246 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db5f30ca-ef98-4f4e-97e6-d3880d6f31a4" containerName="glance-log" Oct 04 03:28:48 crc kubenswrapper[4742]: I1004 03:28:48.579251 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="db5f30ca-ef98-4f4e-97e6-d3880d6f31a4" containerName="glance-log" Oct 04 03:28:48 crc kubenswrapper[4742]: I1004 03:28:48.579384 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="10ce3a20-645d-4c51-9736-87104a27f12f" containerName="glance-log" Oct 04 03:28:48 crc kubenswrapper[4742]: I1004 03:28:48.579396 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="db5f30ca-ef98-4f4e-97e6-d3880d6f31a4" containerName="glance-log" Oct 04 03:28:48 crc kubenswrapper[4742]: I1004 03:28:48.579411 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d7e48d9-5aba-44f1-a31f-fb353d8e034d" containerName="openstackclient" Oct 04 03:28:48 crc kubenswrapper[4742]: I1004 03:28:48.579422 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="db5f30ca-ef98-4f4e-97e6-d3880d6f31a4" containerName="glance-httpd" Oct 04 03:28:48 crc kubenswrapper[4742]: I1004 03:28:48.579429 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="10ce3a20-645d-4c51-9736-87104a27f12f" containerName="glance-httpd" Oct 04 03:28:48 crc kubenswrapper[4742]: I1004 03:28:48.579438 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab7fff03-3721-423c-9a11-8db657c1bde6" containerName="mariadb-account-delete" Oct 04 03:28:48 crc kubenswrapper[4742]: I1004 03:28:48.579865 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-96mj9" Oct 04 03:28:48 crc kubenswrapper[4742]: I1004 03:28:48.590301 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-96mj9"] Oct 04 03:28:48 crc kubenswrapper[4742]: I1004 03:28:48.653106 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmjbb\" (UniqueName: \"kubernetes.io/projected/f04be97b-e81d-4ca3-bdd4-79289f90cf69-kube-api-access-kmjbb\") pod \"glance-db-create-96mj9\" (UID: \"f04be97b-e81d-4ca3-bdd4-79289f90cf69\") " pod="glance-kuttl-tests/glance-db-create-96mj9" Oct 04 03:28:48 crc kubenswrapper[4742]: I1004 03:28:48.757264 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmjbb\" (UniqueName: \"kubernetes.io/projected/f04be97b-e81d-4ca3-bdd4-79289f90cf69-kube-api-access-kmjbb\") pod \"glance-db-create-96mj9\" (UID: \"f04be97b-e81d-4ca3-bdd4-79289f90cf69\") " pod="glance-kuttl-tests/glance-db-create-96mj9" Oct 04 03:28:48 crc kubenswrapper[4742]: I1004 03:28:48.796503 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmjbb\" (UniqueName: \"kubernetes.io/projected/f04be97b-e81d-4ca3-bdd4-79289f90cf69-kube-api-access-kmjbb\") pod \"glance-db-create-96mj9\" (UID: \"f04be97b-e81d-4ca3-bdd4-79289f90cf69\") " pod="glance-kuttl-tests/glance-db-create-96mj9" Oct 04 03:28:48 crc kubenswrapper[4742]: I1004 03:28:48.871822 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10ce3a20-645d-4c51-9736-87104a27f12f" path="/var/lib/kubelet/pods/10ce3a20-645d-4c51-9736-87104a27f12f/volumes" Oct 04 03:28:48 crc kubenswrapper[4742]: I1004 03:28:48.872619 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e2c0638-a70e-4d1f-9cb1-ea9d43f058b4" path="/var/lib/kubelet/pods/8e2c0638-a70e-4d1f-9cb1-ea9d43f058b4/volumes" Oct 04 03:28:48 crc kubenswrapper[4742]: I1004 03:28:48.873605 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fdc4a83-9c85-4de1-b2ba-6bfd995d41c8" path="/var/lib/kubelet/pods/8fdc4a83-9c85-4de1-b2ba-6bfd995d41c8/volumes" Oct 04 03:28:48 crc kubenswrapper[4742]: I1004 03:28:48.875048 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab7fff03-3721-423c-9a11-8db657c1bde6" path="/var/lib/kubelet/pods/ab7fff03-3721-423c-9a11-8db657c1bde6/volumes" Oct 04 03:28:48 crc kubenswrapper[4742]: I1004 03:28:48.875812 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db5f30ca-ef98-4f4e-97e6-d3880d6f31a4" path="/var/lib/kubelet/pods/db5f30ca-ef98-4f4e-97e6-d3880d6f31a4/volumes" Oct 04 03:28:48 crc kubenswrapper[4742]: I1004 03:28:48.906620 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-96mj9" Oct 04 03:28:51 crc kubenswrapper[4742]: I1004 03:28:51.263770 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-96mj9"] Oct 04 03:28:51 crc kubenswrapper[4742]: I1004 03:28:51.832426 4742 generic.go:334] "Generic (PLEG): container finished" podID="f04be97b-e81d-4ca3-bdd4-79289f90cf69" containerID="cd62b165b955c1d17f16d0bbfe0e35424a58f8197e129e653300ccce64a8fa31" exitCode=0 Oct 04 03:28:51 crc kubenswrapper[4742]: I1004 03:28:51.832512 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-96mj9" event={"ID":"f04be97b-e81d-4ca3-bdd4-79289f90cf69","Type":"ContainerDied","Data":"cd62b165b955c1d17f16d0bbfe0e35424a58f8197e129e653300ccce64a8fa31"} Oct 04 03:28:51 crc kubenswrapper[4742]: I1004 03:28:51.832929 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-96mj9" event={"ID":"f04be97b-e81d-4ca3-bdd4-79289f90cf69","Type":"ContainerStarted","Data":"c2e2e529a33bd89ef1f04f0da963c1125967b8728327c5ad0a00418e360699f6"} Oct 04 03:28:53 crc kubenswrapper[4742]: I1004 03:28:53.205015 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-96mj9" Oct 04 03:28:53 crc kubenswrapper[4742]: I1004 03:28:53.369210 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmjbb\" (UniqueName: \"kubernetes.io/projected/f04be97b-e81d-4ca3-bdd4-79289f90cf69-kube-api-access-kmjbb\") pod \"f04be97b-e81d-4ca3-bdd4-79289f90cf69\" (UID: \"f04be97b-e81d-4ca3-bdd4-79289f90cf69\") " Oct 04 03:28:53 crc kubenswrapper[4742]: I1004 03:28:53.379185 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f04be97b-e81d-4ca3-bdd4-79289f90cf69-kube-api-access-kmjbb" (OuterVolumeSpecName: "kube-api-access-kmjbb") pod "f04be97b-e81d-4ca3-bdd4-79289f90cf69" (UID: "f04be97b-e81d-4ca3-bdd4-79289f90cf69"). InnerVolumeSpecName "kube-api-access-kmjbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:28:53 crc kubenswrapper[4742]: I1004 03:28:53.472067 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmjbb\" (UniqueName: \"kubernetes.io/projected/f04be97b-e81d-4ca3-bdd4-79289f90cf69-kube-api-access-kmjbb\") on node \"crc\" DevicePath \"\"" Oct 04 03:28:53 crc kubenswrapper[4742]: I1004 03:28:53.863435 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-96mj9" event={"ID":"f04be97b-e81d-4ca3-bdd4-79289f90cf69","Type":"ContainerDied","Data":"c2e2e529a33bd89ef1f04f0da963c1125967b8728327c5ad0a00418e360699f6"} Oct 04 03:28:53 crc kubenswrapper[4742]: I1004 03:28:53.863479 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-96mj9" Oct 04 03:28:53 crc kubenswrapper[4742]: I1004 03:28:53.863497 4742 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c2e2e529a33bd89ef1f04f0da963c1125967b8728327c5ad0a00418e360699f6" Oct 04 03:29:08 crc kubenswrapper[4742]: I1004 03:29:08.745180 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-34c3-account-create-xx22s"] Oct 04 03:29:08 crc kubenswrapper[4742]: E1004 03:29:08.746699 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f04be97b-e81d-4ca3-bdd4-79289f90cf69" containerName="mariadb-database-create" Oct 04 03:29:08 crc kubenswrapper[4742]: I1004 03:29:08.746729 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="f04be97b-e81d-4ca3-bdd4-79289f90cf69" containerName="mariadb-database-create" Oct 04 03:29:08 crc kubenswrapper[4742]: I1004 03:29:08.747071 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="f04be97b-e81d-4ca3-bdd4-79289f90cf69" containerName="mariadb-database-create" Oct 04 03:29:08 crc kubenswrapper[4742]: I1004 03:29:08.747987 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-34c3-account-create-xx22s" Oct 04 03:29:08 crc kubenswrapper[4742]: I1004 03:29:08.751364 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Oct 04 03:29:08 crc kubenswrapper[4742]: I1004 03:29:08.756186 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-34c3-account-create-xx22s"] Oct 04 03:29:08 crc kubenswrapper[4742]: I1004 03:29:08.886458 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gj4dg\" (UniqueName: \"kubernetes.io/projected/f86185dc-96f0-473c-b26d-a06e54d74ffc-kube-api-access-gj4dg\") pod \"glance-34c3-account-create-xx22s\" (UID: \"f86185dc-96f0-473c-b26d-a06e54d74ffc\") " pod="glance-kuttl-tests/glance-34c3-account-create-xx22s" Oct 04 03:29:08 crc kubenswrapper[4742]: I1004 03:29:08.989746 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gj4dg\" (UniqueName: \"kubernetes.io/projected/f86185dc-96f0-473c-b26d-a06e54d74ffc-kube-api-access-gj4dg\") pod \"glance-34c3-account-create-xx22s\" (UID: \"f86185dc-96f0-473c-b26d-a06e54d74ffc\") " pod="glance-kuttl-tests/glance-34c3-account-create-xx22s" Oct 04 03:29:09 crc kubenswrapper[4742]: I1004 03:29:09.035117 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gj4dg\" (UniqueName: \"kubernetes.io/projected/f86185dc-96f0-473c-b26d-a06e54d74ffc-kube-api-access-gj4dg\") pod \"glance-34c3-account-create-xx22s\" (UID: \"f86185dc-96f0-473c-b26d-a06e54d74ffc\") " pod="glance-kuttl-tests/glance-34c3-account-create-xx22s" Oct 04 03:29:09 crc kubenswrapper[4742]: I1004 03:29:09.091142 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-34c3-account-create-xx22s" Oct 04 03:29:09 crc kubenswrapper[4742]: I1004 03:29:09.402648 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-34c3-account-create-xx22s"] Oct 04 03:29:10 crc kubenswrapper[4742]: I1004 03:29:10.041188 4742 generic.go:334] "Generic (PLEG): container finished" podID="f86185dc-96f0-473c-b26d-a06e54d74ffc" containerID="234ad6f29f016ac8c18cf7f2652ac3632eb27cdad39667dd4043b3cfcc1c4e8a" exitCode=0 Oct 04 03:29:10 crc kubenswrapper[4742]: I1004 03:29:10.041345 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-34c3-account-create-xx22s" event={"ID":"f86185dc-96f0-473c-b26d-a06e54d74ffc","Type":"ContainerDied","Data":"234ad6f29f016ac8c18cf7f2652ac3632eb27cdad39667dd4043b3cfcc1c4e8a"} Oct 04 03:29:10 crc kubenswrapper[4742]: I1004 03:29:10.041890 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-34c3-account-create-xx22s" event={"ID":"f86185dc-96f0-473c-b26d-a06e54d74ffc","Type":"ContainerStarted","Data":"09fb8e2772b1c7187ba7be34bdf29431b2de44160330a7750bca685ab322c767"} Oct 04 03:29:11 crc kubenswrapper[4742]: I1004 03:29:11.444813 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-34c3-account-create-xx22s" Oct 04 03:29:11 crc kubenswrapper[4742]: I1004 03:29:11.567946 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gj4dg\" (UniqueName: \"kubernetes.io/projected/f86185dc-96f0-473c-b26d-a06e54d74ffc-kube-api-access-gj4dg\") pod \"f86185dc-96f0-473c-b26d-a06e54d74ffc\" (UID: \"f86185dc-96f0-473c-b26d-a06e54d74ffc\") " Oct 04 03:29:11 crc kubenswrapper[4742]: I1004 03:29:11.576369 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f86185dc-96f0-473c-b26d-a06e54d74ffc-kube-api-access-gj4dg" (OuterVolumeSpecName: "kube-api-access-gj4dg") pod "f86185dc-96f0-473c-b26d-a06e54d74ffc" (UID: "f86185dc-96f0-473c-b26d-a06e54d74ffc"). InnerVolumeSpecName "kube-api-access-gj4dg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:29:11 crc kubenswrapper[4742]: I1004 03:29:11.669734 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gj4dg\" (UniqueName: \"kubernetes.io/projected/f86185dc-96f0-473c-b26d-a06e54d74ffc-kube-api-access-gj4dg\") on node \"crc\" DevicePath \"\"" Oct 04 03:29:12 crc kubenswrapper[4742]: I1004 03:29:12.080459 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-34c3-account-create-xx22s" event={"ID":"f86185dc-96f0-473c-b26d-a06e54d74ffc","Type":"ContainerDied","Data":"09fb8e2772b1c7187ba7be34bdf29431b2de44160330a7750bca685ab322c767"} Oct 04 03:29:12 crc kubenswrapper[4742]: I1004 03:29:12.080523 4742 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="09fb8e2772b1c7187ba7be34bdf29431b2de44160330a7750bca685ab322c767" Oct 04 03:29:12 crc kubenswrapper[4742]: I1004 03:29:12.080545 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-34c3-account-create-xx22s" Oct 04 03:29:13 crc kubenswrapper[4742]: I1004 03:29:13.817524 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-4l7pq"] Oct 04 03:29:13 crc kubenswrapper[4742]: E1004 03:29:13.818544 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f86185dc-96f0-473c-b26d-a06e54d74ffc" containerName="mariadb-account-create" Oct 04 03:29:13 crc kubenswrapper[4742]: I1004 03:29:13.818565 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="f86185dc-96f0-473c-b26d-a06e54d74ffc" containerName="mariadb-account-create" Oct 04 03:29:13 crc kubenswrapper[4742]: I1004 03:29:13.818768 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="f86185dc-96f0-473c-b26d-a06e54d74ffc" containerName="mariadb-account-create" Oct 04 03:29:13 crc kubenswrapper[4742]: I1004 03:29:13.819833 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-4l7pq" Oct 04 03:29:13 crc kubenswrapper[4742]: I1004 03:29:13.823414 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Oct 04 03:29:13 crc kubenswrapper[4742]: I1004 03:29:13.824613 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-pmsmk" Oct 04 03:29:13 crc kubenswrapper[4742]: I1004 03:29:13.826601 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Oct 04 03:29:13 crc kubenswrapper[4742]: I1004 03:29:13.841026 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-4l7pq"] Oct 04 03:29:13 crc kubenswrapper[4742]: I1004 03:29:13.918755 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbpfr\" (UniqueName: \"kubernetes.io/projected/ce4c0f92-5f63-4b3c-8834-85ebbc047de7-kube-api-access-gbpfr\") pod \"glance-db-sync-4l7pq\" (UID: \"ce4c0f92-5f63-4b3c-8834-85ebbc047de7\") " pod="glance-kuttl-tests/glance-db-sync-4l7pq" Oct 04 03:29:13 crc kubenswrapper[4742]: I1004 03:29:13.918986 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce4c0f92-5f63-4b3c-8834-85ebbc047de7-combined-ca-bundle\") pod \"glance-db-sync-4l7pq\" (UID: \"ce4c0f92-5f63-4b3c-8834-85ebbc047de7\") " pod="glance-kuttl-tests/glance-db-sync-4l7pq" Oct 04 03:29:13 crc kubenswrapper[4742]: I1004 03:29:13.919603 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce4c0f92-5f63-4b3c-8834-85ebbc047de7-config-data\") pod \"glance-db-sync-4l7pq\" (UID: \"ce4c0f92-5f63-4b3c-8834-85ebbc047de7\") " pod="glance-kuttl-tests/glance-db-sync-4l7pq" Oct 04 03:29:13 crc kubenswrapper[4742]: I1004 03:29:13.919751 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ce4c0f92-5f63-4b3c-8834-85ebbc047de7-db-sync-config-data\") pod \"glance-db-sync-4l7pq\" (UID: \"ce4c0f92-5f63-4b3c-8834-85ebbc047de7\") " pod="glance-kuttl-tests/glance-db-sync-4l7pq" Oct 04 03:29:14 crc kubenswrapper[4742]: I1004 03:29:14.022117 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce4c0f92-5f63-4b3c-8834-85ebbc047de7-config-data\") pod \"glance-db-sync-4l7pq\" (UID: \"ce4c0f92-5f63-4b3c-8834-85ebbc047de7\") " pod="glance-kuttl-tests/glance-db-sync-4l7pq" Oct 04 03:29:14 crc kubenswrapper[4742]: I1004 03:29:14.022213 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ce4c0f92-5f63-4b3c-8834-85ebbc047de7-db-sync-config-data\") pod \"glance-db-sync-4l7pq\" (UID: \"ce4c0f92-5f63-4b3c-8834-85ebbc047de7\") " pod="glance-kuttl-tests/glance-db-sync-4l7pq" Oct 04 03:29:14 crc kubenswrapper[4742]: I1004 03:29:14.022308 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbpfr\" (UniqueName: \"kubernetes.io/projected/ce4c0f92-5f63-4b3c-8834-85ebbc047de7-kube-api-access-gbpfr\") pod \"glance-db-sync-4l7pq\" (UID: \"ce4c0f92-5f63-4b3c-8834-85ebbc047de7\") " pod="glance-kuttl-tests/glance-db-sync-4l7pq" Oct 04 03:29:14 crc kubenswrapper[4742]: I1004 03:29:14.022335 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce4c0f92-5f63-4b3c-8834-85ebbc047de7-combined-ca-bundle\") pod \"glance-db-sync-4l7pq\" (UID: \"ce4c0f92-5f63-4b3c-8834-85ebbc047de7\") " pod="glance-kuttl-tests/glance-db-sync-4l7pq" Oct 04 03:29:14 crc kubenswrapper[4742]: I1004 03:29:14.030816 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce4c0f92-5f63-4b3c-8834-85ebbc047de7-config-data\") pod \"glance-db-sync-4l7pq\" (UID: \"ce4c0f92-5f63-4b3c-8834-85ebbc047de7\") " pod="glance-kuttl-tests/glance-db-sync-4l7pq" Oct 04 03:29:14 crc kubenswrapper[4742]: I1004 03:29:14.043148 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce4c0f92-5f63-4b3c-8834-85ebbc047de7-combined-ca-bundle\") pod \"glance-db-sync-4l7pq\" (UID: \"ce4c0f92-5f63-4b3c-8834-85ebbc047de7\") " pod="glance-kuttl-tests/glance-db-sync-4l7pq" Oct 04 03:29:14 crc kubenswrapper[4742]: I1004 03:29:14.043225 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ce4c0f92-5f63-4b3c-8834-85ebbc047de7-db-sync-config-data\") pod \"glance-db-sync-4l7pq\" (UID: \"ce4c0f92-5f63-4b3c-8834-85ebbc047de7\") " pod="glance-kuttl-tests/glance-db-sync-4l7pq" Oct 04 03:29:14 crc kubenswrapper[4742]: I1004 03:29:14.063442 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbpfr\" (UniqueName: \"kubernetes.io/projected/ce4c0f92-5f63-4b3c-8834-85ebbc047de7-kube-api-access-gbpfr\") pod \"glance-db-sync-4l7pq\" (UID: \"ce4c0f92-5f63-4b3c-8834-85ebbc047de7\") " pod="glance-kuttl-tests/glance-db-sync-4l7pq" Oct 04 03:29:14 crc kubenswrapper[4742]: I1004 03:29:14.150479 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-4l7pq" Oct 04 03:29:14 crc kubenswrapper[4742]: I1004 03:29:14.504544 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-4l7pq"] Oct 04 03:29:14 crc kubenswrapper[4742]: W1004 03:29:14.510686 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce4c0f92_5f63_4b3c_8834_85ebbc047de7.slice/crio-73eb2d805012b7225f0405df96307cf770ddd537e8e0017b9e8705ad751a66b0 WatchSource:0}: Error finding container 73eb2d805012b7225f0405df96307cf770ddd537e8e0017b9e8705ad751a66b0: Status 404 returned error can't find the container with id 73eb2d805012b7225f0405df96307cf770ddd537e8e0017b9e8705ad751a66b0 Oct 04 03:29:15 crc kubenswrapper[4742]: I1004 03:29:15.118660 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-4l7pq" event={"ID":"ce4c0f92-5f63-4b3c-8834-85ebbc047de7","Type":"ContainerStarted","Data":"72d7cb4a52e36ca8b35f99369b884ceaf86641e02e37c43d3d81f74ad77f5e46"} Oct 04 03:29:15 crc kubenswrapper[4742]: I1004 03:29:15.119178 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-4l7pq" event={"ID":"ce4c0f92-5f63-4b3c-8834-85ebbc047de7","Type":"ContainerStarted","Data":"73eb2d805012b7225f0405df96307cf770ddd537e8e0017b9e8705ad751a66b0"} Oct 04 03:29:15 crc kubenswrapper[4742]: I1004 03:29:15.142545 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-4l7pq" podStartSLOduration=2.142513057 podStartE2EDuration="2.142513057s" podCreationTimestamp="2025-10-04 03:29:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:29:15.135966789 +0000 UTC m=+1121.068303965" watchObservedRunningTime="2025-10-04 03:29:15.142513057 +0000 UTC m=+1121.074850273" Oct 04 03:29:18 crc kubenswrapper[4742]: I1004 03:29:18.153920 4742 generic.go:334] "Generic (PLEG): container finished" podID="ce4c0f92-5f63-4b3c-8834-85ebbc047de7" containerID="72d7cb4a52e36ca8b35f99369b884ceaf86641e02e37c43d3d81f74ad77f5e46" exitCode=0 Oct 04 03:29:18 crc kubenswrapper[4742]: I1004 03:29:18.154028 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-4l7pq" event={"ID":"ce4c0f92-5f63-4b3c-8834-85ebbc047de7","Type":"ContainerDied","Data":"72d7cb4a52e36ca8b35f99369b884ceaf86641e02e37c43d3d81f74ad77f5e46"} Oct 04 03:29:19 crc kubenswrapper[4742]: I1004 03:29:19.553257 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-4l7pq" Oct 04 03:29:19 crc kubenswrapper[4742]: I1004 03:29:19.633807 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce4c0f92-5f63-4b3c-8834-85ebbc047de7-config-data\") pod \"ce4c0f92-5f63-4b3c-8834-85ebbc047de7\" (UID: \"ce4c0f92-5f63-4b3c-8834-85ebbc047de7\") " Oct 04 03:29:19 crc kubenswrapper[4742]: I1004 03:29:19.633898 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbpfr\" (UniqueName: \"kubernetes.io/projected/ce4c0f92-5f63-4b3c-8834-85ebbc047de7-kube-api-access-gbpfr\") pod \"ce4c0f92-5f63-4b3c-8834-85ebbc047de7\" (UID: \"ce4c0f92-5f63-4b3c-8834-85ebbc047de7\") " Oct 04 03:29:19 crc kubenswrapper[4742]: I1004 03:29:19.633994 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ce4c0f92-5f63-4b3c-8834-85ebbc047de7-db-sync-config-data\") pod \"ce4c0f92-5f63-4b3c-8834-85ebbc047de7\" (UID: \"ce4c0f92-5f63-4b3c-8834-85ebbc047de7\") " Oct 04 03:29:19 crc kubenswrapper[4742]: I1004 03:29:19.634128 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce4c0f92-5f63-4b3c-8834-85ebbc047de7-combined-ca-bundle\") pod \"ce4c0f92-5f63-4b3c-8834-85ebbc047de7\" (UID: \"ce4c0f92-5f63-4b3c-8834-85ebbc047de7\") " Oct 04 03:29:19 crc kubenswrapper[4742]: I1004 03:29:19.642168 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce4c0f92-5f63-4b3c-8834-85ebbc047de7-kube-api-access-gbpfr" (OuterVolumeSpecName: "kube-api-access-gbpfr") pod "ce4c0f92-5f63-4b3c-8834-85ebbc047de7" (UID: "ce4c0f92-5f63-4b3c-8834-85ebbc047de7"). InnerVolumeSpecName "kube-api-access-gbpfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:29:19 crc kubenswrapper[4742]: I1004 03:29:19.644854 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce4c0f92-5f63-4b3c-8834-85ebbc047de7-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "ce4c0f92-5f63-4b3c-8834-85ebbc047de7" (UID: "ce4c0f92-5f63-4b3c-8834-85ebbc047de7"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:29:19 crc kubenswrapper[4742]: I1004 03:29:19.684651 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce4c0f92-5f63-4b3c-8834-85ebbc047de7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ce4c0f92-5f63-4b3c-8834-85ebbc047de7" (UID: "ce4c0f92-5f63-4b3c-8834-85ebbc047de7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:29:19 crc kubenswrapper[4742]: I1004 03:29:19.714527 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce4c0f92-5f63-4b3c-8834-85ebbc047de7-config-data" (OuterVolumeSpecName: "config-data") pod "ce4c0f92-5f63-4b3c-8834-85ebbc047de7" (UID: "ce4c0f92-5f63-4b3c-8834-85ebbc047de7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:29:19 crc kubenswrapper[4742]: I1004 03:29:19.736999 4742 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce4c0f92-5f63-4b3c-8834-85ebbc047de7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:29:19 crc kubenswrapper[4742]: I1004 03:29:19.737053 4742 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce4c0f92-5f63-4b3c-8834-85ebbc047de7-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:29:19 crc kubenswrapper[4742]: I1004 03:29:19.737072 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbpfr\" (UniqueName: \"kubernetes.io/projected/ce4c0f92-5f63-4b3c-8834-85ebbc047de7-kube-api-access-gbpfr\") on node \"crc\" DevicePath \"\"" Oct 04 03:29:19 crc kubenswrapper[4742]: I1004 03:29:19.737093 4742 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ce4c0f92-5f63-4b3c-8834-85ebbc047de7-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.175439 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-4l7pq" event={"ID":"ce4c0f92-5f63-4b3c-8834-85ebbc047de7","Type":"ContainerDied","Data":"73eb2d805012b7225f0405df96307cf770ddd537e8e0017b9e8705ad751a66b0"} Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.175489 4742 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73eb2d805012b7225f0405df96307cf770ddd537e8e0017b9e8705ad751a66b0" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.175510 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-4l7pq" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.531350 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 03:29:20 crc kubenswrapper[4742]: E1004 03:29:20.531772 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce4c0f92-5f63-4b3c-8834-85ebbc047de7" containerName="glance-db-sync" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.531860 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce4c0f92-5f63-4b3c-8834-85ebbc047de7" containerName="glance-db-sync" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.532080 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce4c0f92-5f63-4b3c-8834-85ebbc047de7" containerName="glance-db-sync" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.533292 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.535347 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-pmsmk" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.536317 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.539700 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.539776 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.539835 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-internal-svc" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.540207 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-public-svc" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.543952 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 03:29:20 crc kubenswrapper[4742]: E1004 03:29:20.550592 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle config-data glance httpd-run internal-tls-certs kube-api-access-vxzpf logs public-tls-certs scripts], unattached volumes=[], failed to process volumes=[combined-ca-bundle config-data glance httpd-run internal-tls-certs kube-api-access-vxzpf logs public-tls-certs scripts]: context canceled" pod="glance-kuttl-tests/glance-default-single-0" podUID="05547c76-8358-4673-955c-cd59c4864f4e" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.601512 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.651100 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/05547c76-8358-4673-955c-cd59c4864f4e-httpd-run\") pod \"glance-default-single-0\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.651149 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05547c76-8358-4673-955c-cd59c4864f4e-scripts\") pod \"glance-default-single-0\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.651223 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05547c76-8358-4673-955c-cd59c4864f4e-logs\") pod \"glance-default-single-0\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.651255 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/05547c76-8358-4673-955c-cd59c4864f4e-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.651474 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/05547c76-8358-4673-955c-cd59c4864f4e-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.651560 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.651663 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05547c76-8358-4673-955c-cd59c4864f4e-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.651702 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05547c76-8358-4673-955c-cd59c4864f4e-config-data\") pod \"glance-default-single-0\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.651854 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxzpf\" (UniqueName: \"kubernetes.io/projected/05547c76-8358-4673-955c-cd59c4864f4e-kube-api-access-vxzpf\") pod \"glance-default-single-0\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.753127 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/05547c76-8358-4673-955c-cd59c4864f4e-httpd-run\") pod \"glance-default-single-0\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.753179 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05547c76-8358-4673-955c-cd59c4864f4e-scripts\") pod \"glance-default-single-0\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.753230 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05547c76-8358-4673-955c-cd59c4864f4e-logs\") pod \"glance-default-single-0\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.753265 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/05547c76-8358-4673-955c-cd59c4864f4e-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.753340 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/05547c76-8358-4673-955c-cd59c4864f4e-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.753368 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.753390 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05547c76-8358-4673-955c-cd59c4864f4e-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.753414 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05547c76-8358-4673-955c-cd59c4864f4e-config-data\") pod \"glance-default-single-0\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.753457 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxzpf\" (UniqueName: \"kubernetes.io/projected/05547c76-8358-4673-955c-cd59c4864f4e-kube-api-access-vxzpf\") pod \"glance-default-single-0\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.754042 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05547c76-8358-4673-955c-cd59c4864f4e-logs\") pod \"glance-default-single-0\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.754716 4742 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.759618 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/05547c76-8358-4673-955c-cd59c4864f4e-httpd-run\") pod \"glance-default-single-0\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.760496 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05547c76-8358-4673-955c-cd59c4864f4e-scripts\") pod \"glance-default-single-0\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.761056 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05547c76-8358-4673-955c-cd59c4864f4e-config-data\") pod \"glance-default-single-0\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.762244 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/05547c76-8358-4673-955c-cd59c4864f4e-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.762889 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/05547c76-8358-4673-955c-cd59c4864f4e-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.765891 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05547c76-8358-4673-955c-cd59c4864f4e-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.776856 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxzpf\" (UniqueName: \"kubernetes.io/projected/05547c76-8358-4673-955c-cd59c4864f4e-kube-api-access-vxzpf\") pod \"glance-default-single-0\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:20 crc kubenswrapper[4742]: I1004 03:29:20.778672 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:21 crc kubenswrapper[4742]: I1004 03:29:21.184873 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:21 crc kubenswrapper[4742]: I1004 03:29:21.200301 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:21 crc kubenswrapper[4742]: I1004 03:29:21.260813 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05547c76-8358-4673-955c-cd59c4864f4e-scripts\") pod \"05547c76-8358-4673-955c-cd59c4864f4e\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") " Oct 04 03:29:21 crc kubenswrapper[4742]: I1004 03:29:21.260883 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"05547c76-8358-4673-955c-cd59c4864f4e\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") " Oct 04 03:29:21 crc kubenswrapper[4742]: I1004 03:29:21.260916 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05547c76-8358-4673-955c-cd59c4864f4e-logs\") pod \"05547c76-8358-4673-955c-cd59c4864f4e\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") " Oct 04 03:29:21 crc kubenswrapper[4742]: I1004 03:29:21.260954 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/05547c76-8358-4673-955c-cd59c4864f4e-public-tls-certs\") pod \"05547c76-8358-4673-955c-cd59c4864f4e\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") " Oct 04 03:29:21 crc kubenswrapper[4742]: I1004 03:29:21.260978 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxzpf\" (UniqueName: \"kubernetes.io/projected/05547c76-8358-4673-955c-cd59c4864f4e-kube-api-access-vxzpf\") pod \"05547c76-8358-4673-955c-cd59c4864f4e\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") " Oct 04 03:29:21 crc kubenswrapper[4742]: I1004 03:29:21.261016 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05547c76-8358-4673-955c-cd59c4864f4e-combined-ca-bundle\") pod \"05547c76-8358-4673-955c-cd59c4864f4e\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") " Oct 04 03:29:21 crc kubenswrapper[4742]: I1004 03:29:21.261047 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/05547c76-8358-4673-955c-cd59c4864f4e-httpd-run\") pod \"05547c76-8358-4673-955c-cd59c4864f4e\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") " Oct 04 03:29:21 crc kubenswrapper[4742]: I1004 03:29:21.261073 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05547c76-8358-4673-955c-cd59c4864f4e-config-data\") pod \"05547c76-8358-4673-955c-cd59c4864f4e\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") " Oct 04 03:29:21 crc kubenswrapper[4742]: I1004 03:29:21.261125 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/05547c76-8358-4673-955c-cd59c4864f4e-internal-tls-certs\") pod \"05547c76-8358-4673-955c-cd59c4864f4e\" (UID: \"05547c76-8358-4673-955c-cd59c4864f4e\") " Oct 04 03:29:21 crc kubenswrapper[4742]: I1004 03:29:21.262059 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05547c76-8358-4673-955c-cd59c4864f4e-logs" (OuterVolumeSpecName: "logs") pod "05547c76-8358-4673-955c-cd59c4864f4e" (UID: "05547c76-8358-4673-955c-cd59c4864f4e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:29:21 crc kubenswrapper[4742]: I1004 03:29:21.262929 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05547c76-8358-4673-955c-cd59c4864f4e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "05547c76-8358-4673-955c-cd59c4864f4e" (UID: "05547c76-8358-4673-955c-cd59c4864f4e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:29:21 crc kubenswrapper[4742]: I1004 03:29:21.265731 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05547c76-8358-4673-955c-cd59c4864f4e-scripts" (OuterVolumeSpecName: "scripts") pod "05547c76-8358-4673-955c-cd59c4864f4e" (UID: "05547c76-8358-4673-955c-cd59c4864f4e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:29:21 crc kubenswrapper[4742]: I1004 03:29:21.265789 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "05547c76-8358-4673-955c-cd59c4864f4e" (UID: "05547c76-8358-4673-955c-cd59c4864f4e"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 03:29:21 crc kubenswrapper[4742]: I1004 03:29:21.266306 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05547c76-8358-4673-955c-cd59c4864f4e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "05547c76-8358-4673-955c-cd59c4864f4e" (UID: "05547c76-8358-4673-955c-cd59c4864f4e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:29:21 crc kubenswrapper[4742]: I1004 03:29:21.267157 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05547c76-8358-4673-955c-cd59c4864f4e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "05547c76-8358-4673-955c-cd59c4864f4e" (UID: "05547c76-8358-4673-955c-cd59c4864f4e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:29:21 crc kubenswrapper[4742]: I1004 03:29:21.267990 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05547c76-8358-4673-955c-cd59c4864f4e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "05547c76-8358-4673-955c-cd59c4864f4e" (UID: "05547c76-8358-4673-955c-cd59c4864f4e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:29:21 crc kubenswrapper[4742]: I1004 03:29:21.269032 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05547c76-8358-4673-955c-cd59c4864f4e-kube-api-access-vxzpf" (OuterVolumeSpecName: "kube-api-access-vxzpf") pod "05547c76-8358-4673-955c-cd59c4864f4e" (UID: "05547c76-8358-4673-955c-cd59c4864f4e"). InnerVolumeSpecName "kube-api-access-vxzpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:29:21 crc kubenswrapper[4742]: I1004 03:29:21.280550 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05547c76-8358-4673-955c-cd59c4864f4e-config-data" (OuterVolumeSpecName: "config-data") pod "05547c76-8358-4673-955c-cd59c4864f4e" (UID: "05547c76-8358-4673-955c-cd59c4864f4e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:29:21 crc kubenswrapper[4742]: I1004 03:29:21.362784 4742 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05547c76-8358-4673-955c-cd59c4864f4e-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:29:21 crc kubenswrapper[4742]: I1004 03:29:21.362829 4742 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/05547c76-8358-4673-955c-cd59c4864f4e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:29:21 crc kubenswrapper[4742]: I1004 03:29:21.362845 4742 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05547c76-8358-4673-955c-cd59c4864f4e-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:29:21 crc kubenswrapper[4742]: I1004 03:29:21.362886 4742 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 04 03:29:21 crc kubenswrapper[4742]: I1004 03:29:21.362908 4742 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05547c76-8358-4673-955c-cd59c4864f4e-logs\") on node \"crc\" DevicePath \"\"" Oct 04 03:29:21 crc kubenswrapper[4742]: I1004 03:29:21.362921 4742 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/05547c76-8358-4673-955c-cd59c4864f4e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:29:21 crc kubenswrapper[4742]: I1004 03:29:21.362933 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxzpf\" (UniqueName: \"kubernetes.io/projected/05547c76-8358-4673-955c-cd59c4864f4e-kube-api-access-vxzpf\") on node \"crc\" DevicePath \"\"" Oct 04 03:29:21 crc kubenswrapper[4742]: I1004 03:29:21.362945 4742 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05547c76-8358-4673-955c-cd59c4864f4e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:29:21 crc kubenswrapper[4742]: I1004 03:29:21.362957 4742 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/05547c76-8358-4673-955c-cd59c4864f4e-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 03:29:21 crc kubenswrapper[4742]: I1004 03:29:21.379391 4742 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 04 03:29:21 crc kubenswrapper[4742]: I1004 03:29:21.464268 4742 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 04 03:29:22 crc kubenswrapper[4742]: I1004 03:29:22.191847 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:22 crc kubenswrapper[4742]: I1004 03:29:22.246612 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 03:29:22 crc kubenswrapper[4742]: I1004 03:29:22.251416 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 03:29:22 crc kubenswrapper[4742]: I1004 03:29:22.873657 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05547c76-8358-4673-955c-cd59c4864f4e" path="/var/lib/kubelet/pods/05547c76-8358-4673-955c-cd59c4864f4e/volumes" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.345304 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.347201 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.351661 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-internal-svc" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.351661 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-public-svc" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.352400 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.353949 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-pmsmk" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.354847 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.362236 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.395203 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.396490 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-logs\") pod \"glance-default-single-0\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.396575 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mztqb\" (UniqueName: \"kubernetes.io/projected/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-kube-api-access-mztqb\") pod \"glance-default-single-0\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.396658 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.396706 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-config-data\") pod \"glance-default-single-0\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.396765 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-scripts\") pod \"glance-default-single-0\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.396812 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.396916 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-httpd-run\") pod \"glance-default-single-0\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.396981 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.397054 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.498622 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.498733 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-logs\") pod \"glance-default-single-0\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.498774 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mztqb\" (UniqueName: \"kubernetes.io/projected/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-kube-api-access-mztqb\") pod \"glance-default-single-0\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.498825 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.498861 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-config-data\") pod \"glance-default-single-0\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.498911 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-scripts\") pod \"glance-default-single-0\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.498951 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.499043 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-httpd-run\") pod \"glance-default-single-0\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.499106 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.499576 4742 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.500232 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-logs\") pod \"glance-default-single-0\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.500668 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-httpd-run\") pod \"glance-default-single-0\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.509611 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-config-data\") pod \"glance-default-single-0\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.514403 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-scripts\") pod \"glance-default-single-0\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.517216 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.517498 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.524500 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.528189 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mztqb\" (UniqueName: \"kubernetes.io/projected/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-kube-api-access-mztqb\") pod \"glance-default-single-0\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.549148 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.685982 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:23 crc kubenswrapper[4742]: I1004 03:29:23.987922 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 03:29:24 crc kubenswrapper[4742]: I1004 03:29:24.214025 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9","Type":"ContainerStarted","Data":"2b143015cf146486649564b002bfb8474d461843e390afa361ea1945e2be588a"} Oct 04 03:29:25 crc kubenswrapper[4742]: I1004 03:29:25.235105 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9","Type":"ContainerStarted","Data":"692eb0d8844f11c1234e6ef786f3adc2dd0d75670c5a90f3e69549828d1aab49"} Oct 04 03:29:26 crc kubenswrapper[4742]: I1004 03:29:26.254894 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9","Type":"ContainerStarted","Data":"0a84ae9bca27c8099731f069db571b62918cb69b0af7d1cc74179d6ad39c8351"} Oct 04 03:29:26 crc kubenswrapper[4742]: I1004 03:29:26.296715 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=3.296680769 podStartE2EDuration="3.296680769s" podCreationTimestamp="2025-10-04 03:29:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:29:26.284291431 +0000 UTC m=+1132.216628617" watchObservedRunningTime="2025-10-04 03:29:26.296680769 +0000 UTC m=+1132.229017985" Oct 04 03:29:33 crc kubenswrapper[4742]: I1004 03:29:33.686323 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:33 crc kubenswrapper[4742]: I1004 03:29:33.687435 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:33 crc kubenswrapper[4742]: I1004 03:29:33.722200 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:33 crc kubenswrapper[4742]: I1004 03:29:33.736161 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:34 crc kubenswrapper[4742]: I1004 03:29:34.351518 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:34 crc kubenswrapper[4742]: I1004 03:29:34.352051 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:36 crc kubenswrapper[4742]: I1004 03:29:36.244231 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:36 crc kubenswrapper[4742]: I1004 03:29:36.246157 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:37 crc kubenswrapper[4742]: I1004 03:29:37.277572 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-4l7pq"] Oct 04 03:29:37 crc kubenswrapper[4742]: I1004 03:29:37.287476 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-4l7pq"] Oct 04 03:29:37 crc kubenswrapper[4742]: I1004 03:29:37.376321 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance34c3-account-delete-c2cf8"] Oct 04 03:29:37 crc kubenswrapper[4742]: I1004 03:29:37.377203 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance34c3-account-delete-c2cf8" Oct 04 03:29:37 crc kubenswrapper[4742]: I1004 03:29:37.385566 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 03:29:37 crc kubenswrapper[4742]: I1004 03:29:37.388568 4742 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="glance-kuttl-tests/glance-default-single-0" secret="" err="secret \"glance-glance-dockercfg-pmsmk\" not found" Oct 04 03:29:37 crc kubenswrapper[4742]: I1004 03:29:37.399995 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance34c3-account-delete-c2cf8"] Oct 04 03:29:37 crc kubenswrapper[4742]: I1004 03:29:37.437037 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-96mj9"] Oct 04 03:29:37 crc kubenswrapper[4742]: I1004 03:29:37.441635 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-96mj9"] Oct 04 03:29:37 crc kubenswrapper[4742]: I1004 03:29:37.455875 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvrtm\" (UniqueName: \"kubernetes.io/projected/3954a32b-cd91-49ad-9613-fbe35271edcf-kube-api-access-tvrtm\") pod \"glance34c3-account-delete-c2cf8\" (UID: \"3954a32b-cd91-49ad-9613-fbe35271edcf\") " pod="glance-kuttl-tests/glance34c3-account-delete-c2cf8" Oct 04 03:29:37 crc kubenswrapper[4742]: E1004 03:29:37.456706 4742 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-default-single-config-data: secret "glance-default-single-config-data" not found Oct 04 03:29:37 crc kubenswrapper[4742]: E1004 03:29:37.456759 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-config-data podName:1fd1e584-06b3-4573-8ae5-cdcd9ed004e9 nodeName:}" failed. No retries permitted until 2025-10-04 03:29:37.956741204 +0000 UTC m=+1143.889078380 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-config-data") pod "glance-default-single-0" (UID: "1fd1e584-06b3-4573-8ae5-cdcd9ed004e9") : secret "glance-default-single-config-data" not found Oct 04 03:29:37 crc kubenswrapper[4742]: E1004 03:29:37.457099 4742 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-scripts: secret "glance-scripts" not found Oct 04 03:29:37 crc kubenswrapper[4742]: E1004 03:29:37.457251 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-scripts podName:1fd1e584-06b3-4573-8ae5-cdcd9ed004e9 nodeName:}" failed. No retries permitted until 2025-10-04 03:29:37.957215447 +0000 UTC m=+1143.889552623 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-scripts") pod "glance-default-single-0" (UID: "1fd1e584-06b3-4573-8ae5-cdcd9ed004e9") : secret "glance-scripts" not found Oct 04 03:29:37 crc kubenswrapper[4742]: I1004 03:29:37.462836 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-34c3-account-create-xx22s"] Oct 04 03:29:37 crc kubenswrapper[4742]: I1004 03:29:37.468364 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance34c3-account-delete-c2cf8"] Oct 04 03:29:37 crc kubenswrapper[4742]: E1004 03:29:37.468826 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-tvrtm], unattached volumes=[], failed to process volumes=[]: context canceled" pod="glance-kuttl-tests/glance34c3-account-delete-c2cf8" podUID="3954a32b-cd91-49ad-9613-fbe35271edcf" Oct 04 03:29:37 crc kubenswrapper[4742]: I1004 03:29:37.479355 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-34c3-account-create-xx22s"] Oct 04 03:29:37 crc kubenswrapper[4742]: I1004 03:29:37.557326 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvrtm\" (UniqueName: \"kubernetes.io/projected/3954a32b-cd91-49ad-9613-fbe35271edcf-kube-api-access-tvrtm\") pod \"glance34c3-account-delete-c2cf8\" (UID: \"3954a32b-cd91-49ad-9613-fbe35271edcf\") " pod="glance-kuttl-tests/glance34c3-account-delete-c2cf8" Oct 04 03:29:37 crc kubenswrapper[4742]: I1004 03:29:37.577464 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvrtm\" (UniqueName: \"kubernetes.io/projected/3954a32b-cd91-49ad-9613-fbe35271edcf-kube-api-access-tvrtm\") pod \"glance34c3-account-delete-c2cf8\" (UID: \"3954a32b-cd91-49ad-9613-fbe35271edcf\") " pod="glance-kuttl-tests/glance34c3-account-delete-c2cf8" Oct 04 03:29:37 crc kubenswrapper[4742]: E1004 03:29:37.963667 4742 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-default-single-config-data: secret "glance-default-single-config-data" not found Oct 04 03:29:37 crc kubenswrapper[4742]: E1004 03:29:37.964198 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-config-data podName:1fd1e584-06b3-4573-8ae5-cdcd9ed004e9 nodeName:}" failed. No retries permitted until 2025-10-04 03:29:38.964177726 +0000 UTC m=+1144.896514912 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-config-data") pod "glance-default-single-0" (UID: "1fd1e584-06b3-4573-8ae5-cdcd9ed004e9") : secret "glance-default-single-config-data" not found Oct 04 03:29:37 crc kubenswrapper[4742]: E1004 03:29:37.964344 4742 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-scripts: secret "glance-scripts" not found Oct 04 03:29:37 crc kubenswrapper[4742]: E1004 03:29:37.964376 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-scripts podName:1fd1e584-06b3-4573-8ae5-cdcd9ed004e9 nodeName:}" failed. No retries permitted until 2025-10-04 03:29:38.964367741 +0000 UTC m=+1144.896704927 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-scripts") pod "glance-default-single-0" (UID: "1fd1e584-06b3-4573-8ae5-cdcd9ed004e9") : secret "glance-scripts" not found Oct 04 03:29:38 crc kubenswrapper[4742]: I1004 03:29:38.396187 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance34c3-account-delete-c2cf8" Oct 04 03:29:38 crc kubenswrapper[4742]: I1004 03:29:38.396838 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="1fd1e584-06b3-4573-8ae5-cdcd9ed004e9" containerName="glance-log" containerID="cri-o://692eb0d8844f11c1234e6ef786f3adc2dd0d75670c5a90f3e69549828d1aab49" gracePeriod=30 Oct 04 03:29:38 crc kubenswrapper[4742]: I1004 03:29:38.397377 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="1fd1e584-06b3-4573-8ae5-cdcd9ed004e9" containerName="glance-httpd" containerID="cri-o://0a84ae9bca27c8099731f069db571b62918cb69b0af7d1cc74179d6ad39c8351" gracePeriod=30 Oct 04 03:29:38 crc kubenswrapper[4742]: I1004 03:29:38.408386 4742 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="1fd1e584-06b3-4573-8ae5-cdcd9ed004e9" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.112:9292/healthcheck\": EOF" Oct 04 03:29:38 crc kubenswrapper[4742]: I1004 03:29:38.411032 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance34c3-account-delete-c2cf8" Oct 04 03:29:38 crc kubenswrapper[4742]: I1004 03:29:38.472248 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvrtm\" (UniqueName: \"kubernetes.io/projected/3954a32b-cd91-49ad-9613-fbe35271edcf-kube-api-access-tvrtm\") pod \"3954a32b-cd91-49ad-9613-fbe35271edcf\" (UID: \"3954a32b-cd91-49ad-9613-fbe35271edcf\") " Oct 04 03:29:38 crc kubenswrapper[4742]: I1004 03:29:38.478902 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3954a32b-cd91-49ad-9613-fbe35271edcf-kube-api-access-tvrtm" (OuterVolumeSpecName: "kube-api-access-tvrtm") pod "3954a32b-cd91-49ad-9613-fbe35271edcf" (UID: "3954a32b-cd91-49ad-9613-fbe35271edcf"). InnerVolumeSpecName "kube-api-access-tvrtm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:29:38 crc kubenswrapper[4742]: I1004 03:29:38.574871 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvrtm\" (UniqueName: \"kubernetes.io/projected/3954a32b-cd91-49ad-9613-fbe35271edcf-kube-api-access-tvrtm\") on node \"crc\" DevicePath \"\"" Oct 04 03:29:38 crc kubenswrapper[4742]: I1004 03:29:38.876534 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce4c0f92-5f63-4b3c-8834-85ebbc047de7" path="/var/lib/kubelet/pods/ce4c0f92-5f63-4b3c-8834-85ebbc047de7/volumes" Oct 04 03:29:38 crc kubenswrapper[4742]: I1004 03:29:38.878322 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f04be97b-e81d-4ca3-bdd4-79289f90cf69" path="/var/lib/kubelet/pods/f04be97b-e81d-4ca3-bdd4-79289f90cf69/volumes" Oct 04 03:29:38 crc kubenswrapper[4742]: I1004 03:29:38.878962 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f86185dc-96f0-473c-b26d-a06e54d74ffc" path="/var/lib/kubelet/pods/f86185dc-96f0-473c-b26d-a06e54d74ffc/volumes" Oct 04 03:29:38 crc kubenswrapper[4742]: E1004 03:29:38.982128 4742 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-default-single-config-data: secret "glance-default-single-config-data" not found Oct 04 03:29:38 crc kubenswrapper[4742]: E1004 03:29:38.982263 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-config-data podName:1fd1e584-06b3-4573-8ae5-cdcd9ed004e9 nodeName:}" failed. No retries permitted until 2025-10-04 03:29:40.982232493 +0000 UTC m=+1146.914569709 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-config-data") pod "glance-default-single-0" (UID: "1fd1e584-06b3-4573-8ae5-cdcd9ed004e9") : secret "glance-default-single-config-data" not found Oct 04 03:29:38 crc kubenswrapper[4742]: E1004 03:29:38.982588 4742 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-scripts: secret "glance-scripts" not found Oct 04 03:29:38 crc kubenswrapper[4742]: E1004 03:29:38.982686 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-scripts podName:1fd1e584-06b3-4573-8ae5-cdcd9ed004e9 nodeName:}" failed. No retries permitted until 2025-10-04 03:29:40.982659474 +0000 UTC m=+1146.914996640 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-scripts") pod "glance-default-single-0" (UID: "1fd1e584-06b3-4573-8ae5-cdcd9ed004e9") : secret "glance-scripts" not found Oct 04 03:29:39 crc kubenswrapper[4742]: I1004 03:29:39.419286 4742 generic.go:334] "Generic (PLEG): container finished" podID="1fd1e584-06b3-4573-8ae5-cdcd9ed004e9" containerID="692eb0d8844f11c1234e6ef786f3adc2dd0d75670c5a90f3e69549828d1aab49" exitCode=143 Oct 04 03:29:39 crc kubenswrapper[4742]: I1004 03:29:39.419312 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9","Type":"ContainerDied","Data":"692eb0d8844f11c1234e6ef786f3adc2dd0d75670c5a90f3e69549828d1aab49"} Oct 04 03:29:39 crc kubenswrapper[4742]: I1004 03:29:39.419819 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance34c3-account-delete-c2cf8" Oct 04 03:29:39 crc kubenswrapper[4742]: I1004 03:29:39.458005 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance34c3-account-delete-c2cf8"] Oct 04 03:29:39 crc kubenswrapper[4742]: I1004 03:29:39.465511 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance34c3-account-delete-c2cf8"] Oct 04 03:29:40 crc kubenswrapper[4742]: I1004 03:29:40.869354 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3954a32b-cd91-49ad-9613-fbe35271edcf" path="/var/lib/kubelet/pods/3954a32b-cd91-49ad-9613-fbe35271edcf/volumes" Oct 04 03:29:41 crc kubenswrapper[4742]: E1004 03:29:41.014604 4742 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-scripts: secret "glance-scripts" not found Oct 04 03:29:41 crc kubenswrapper[4742]: E1004 03:29:41.014743 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-scripts podName:1fd1e584-06b3-4573-8ae5-cdcd9ed004e9 nodeName:}" failed. No retries permitted until 2025-10-04 03:29:45.014711331 +0000 UTC m=+1150.947048557 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-scripts") pod "glance-default-single-0" (UID: "1fd1e584-06b3-4573-8ae5-cdcd9ed004e9") : secret "glance-scripts" not found Oct 04 03:29:41 crc kubenswrapper[4742]: E1004 03:29:41.014602 4742 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-default-single-config-data: secret "glance-default-single-config-data" not found Oct 04 03:29:41 crc kubenswrapper[4742]: E1004 03:29:41.014941 4742 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-config-data podName:1fd1e584-06b3-4573-8ae5-cdcd9ed004e9 nodeName:}" failed. No retries permitted until 2025-10-04 03:29:45.014901615 +0000 UTC m=+1150.947238821 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-config-data") pod "glance-default-single-0" (UID: "1fd1e584-06b3-4573-8ae5-cdcd9ed004e9") : secret "glance-default-single-config-data" not found Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.058873 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.131930 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") " Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.132035 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-config-data\") pod \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") " Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.132085 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-scripts\") pod \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") " Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.132140 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-internal-tls-certs\") pod \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") " Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.132181 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-combined-ca-bundle\") pod \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") " Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.132224 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-httpd-run\") pod \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") " Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.132326 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mztqb\" (UniqueName: \"kubernetes.io/projected/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-kube-api-access-mztqb\") pod \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") " Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.132377 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-public-tls-certs\") pod \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") " Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.132454 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-logs\") pod \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\" (UID: \"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9\") " Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.133432 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1fd1e584-06b3-4573-8ae5-cdcd9ed004e9" (UID: "1fd1e584-06b3-4573-8ae5-cdcd9ed004e9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.133469 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-logs" (OuterVolumeSpecName: "logs") pod "1fd1e584-06b3-4573-8ae5-cdcd9ed004e9" (UID: "1fd1e584-06b3-4573-8ae5-cdcd9ed004e9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.134036 4742 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.134095 4742 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-logs\") on node \"crc\" DevicePath \"\"" Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.142500 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "1fd1e584-06b3-4573-8ae5-cdcd9ed004e9" (UID: "1fd1e584-06b3-4573-8ae5-cdcd9ed004e9"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.143140 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-scripts" (OuterVolumeSpecName: "scripts") pod "1fd1e584-06b3-4573-8ae5-cdcd9ed004e9" (UID: "1fd1e584-06b3-4573-8ae5-cdcd9ed004e9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.146547 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-kube-api-access-mztqb" (OuterVolumeSpecName: "kube-api-access-mztqb") pod "1fd1e584-06b3-4573-8ae5-cdcd9ed004e9" (UID: "1fd1e584-06b3-4573-8ae5-cdcd9ed004e9"). InnerVolumeSpecName "kube-api-access-mztqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.187008 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1fd1e584-06b3-4573-8ae5-cdcd9ed004e9" (UID: "1fd1e584-06b3-4573-8ae5-cdcd9ed004e9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.198859 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-config-data" (OuterVolumeSpecName: "config-data") pod "1fd1e584-06b3-4573-8ae5-cdcd9ed004e9" (UID: "1fd1e584-06b3-4573-8ae5-cdcd9ed004e9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.202589 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "1fd1e584-06b3-4573-8ae5-cdcd9ed004e9" (UID: "1fd1e584-06b3-4573-8ae5-cdcd9ed004e9"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.219116 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "1fd1e584-06b3-4573-8ae5-cdcd9ed004e9" (UID: "1fd1e584-06b3-4573-8ae5-cdcd9ed004e9"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.237578 4742 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.237998 4742 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.238182 4742 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.238393 4742 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.238532 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mztqb\" (UniqueName: \"kubernetes.io/projected/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-kube-api-access-mztqb\") on node \"crc\" DevicePath \"\"" Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.238687 4742 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.238875 4742 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.266767 4742 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.340914 4742 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.450731 4742 generic.go:334] "Generic (PLEG): container finished" podID="1fd1e584-06b3-4573-8ae5-cdcd9ed004e9" containerID="0a84ae9bca27c8099731f069db571b62918cb69b0af7d1cc74179d6ad39c8351" exitCode=0 Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.450811 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9","Type":"ContainerDied","Data":"0a84ae9bca27c8099731f069db571b62918cb69b0af7d1cc74179d6ad39c8351"} Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.450849 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"1fd1e584-06b3-4573-8ae5-cdcd9ed004e9","Type":"ContainerDied","Data":"2b143015cf146486649564b002bfb8474d461843e390afa361ea1945e2be588a"} Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.450898 4742 scope.go:117] "RemoveContainer" containerID="0a84ae9bca27c8099731f069db571b62918cb69b0af7d1cc74179d6ad39c8351" Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.450930 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.486781 4742 scope.go:117] "RemoveContainer" containerID="692eb0d8844f11c1234e6ef786f3adc2dd0d75670c5a90f3e69549828d1aab49" Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.487034 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.493395 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.510568 4742 scope.go:117] "RemoveContainer" containerID="0a84ae9bca27c8099731f069db571b62918cb69b0af7d1cc74179d6ad39c8351" Oct 04 03:29:42 crc kubenswrapper[4742]: E1004 03:29:42.511141 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a84ae9bca27c8099731f069db571b62918cb69b0af7d1cc74179d6ad39c8351\": container with ID starting with 0a84ae9bca27c8099731f069db571b62918cb69b0af7d1cc74179d6ad39c8351 not found: ID does not exist" containerID="0a84ae9bca27c8099731f069db571b62918cb69b0af7d1cc74179d6ad39c8351" Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.511208 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a84ae9bca27c8099731f069db571b62918cb69b0af7d1cc74179d6ad39c8351"} err="failed to get container status \"0a84ae9bca27c8099731f069db571b62918cb69b0af7d1cc74179d6ad39c8351\": rpc error: code = NotFound desc = could not find container \"0a84ae9bca27c8099731f069db571b62918cb69b0af7d1cc74179d6ad39c8351\": container with ID starting with 0a84ae9bca27c8099731f069db571b62918cb69b0af7d1cc74179d6ad39c8351 not found: ID does not exist" Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.511301 4742 scope.go:117] "RemoveContainer" containerID="692eb0d8844f11c1234e6ef786f3adc2dd0d75670c5a90f3e69549828d1aab49" Oct 04 03:29:42 crc kubenswrapper[4742]: E1004 03:29:42.511919 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"692eb0d8844f11c1234e6ef786f3adc2dd0d75670c5a90f3e69549828d1aab49\": container with ID starting with 692eb0d8844f11c1234e6ef786f3adc2dd0d75670c5a90f3e69549828d1aab49 not found: ID does not exist" containerID="692eb0d8844f11c1234e6ef786f3adc2dd0d75670c5a90f3e69549828d1aab49" Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.511995 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"692eb0d8844f11c1234e6ef786f3adc2dd0d75670c5a90f3e69549828d1aab49"} err="failed to get container status \"692eb0d8844f11c1234e6ef786f3adc2dd0d75670c5a90f3e69549828d1aab49\": rpc error: code = NotFound desc = could not find container \"692eb0d8844f11c1234e6ef786f3adc2dd0d75670c5a90f3e69549828d1aab49\": container with ID starting with 692eb0d8844f11c1234e6ef786f3adc2dd0d75670c5a90f3e69549828d1aab49 not found: ID does not exist" Oct 04 03:29:42 crc kubenswrapper[4742]: I1004 03:29:42.871006 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fd1e584-06b3-4573-8ae5-cdcd9ed004e9" path="/var/lib/kubelet/pods/1fd1e584-06b3-4573-8ae5-cdcd9ed004e9/volumes" Oct 04 03:29:44 crc kubenswrapper[4742]: I1004 03:29:44.265093 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-cm85w"] Oct 04 03:29:44 crc kubenswrapper[4742]: E1004 03:29:44.265892 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fd1e584-06b3-4573-8ae5-cdcd9ed004e9" containerName="glance-httpd" Oct 04 03:29:44 crc kubenswrapper[4742]: I1004 03:29:44.265910 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fd1e584-06b3-4573-8ae5-cdcd9ed004e9" containerName="glance-httpd" Oct 04 03:29:44 crc kubenswrapper[4742]: E1004 03:29:44.265926 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fd1e584-06b3-4573-8ae5-cdcd9ed004e9" containerName="glance-log" Oct 04 03:29:44 crc kubenswrapper[4742]: I1004 03:29:44.265933 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fd1e584-06b3-4573-8ae5-cdcd9ed004e9" containerName="glance-log" Oct 04 03:29:44 crc kubenswrapper[4742]: I1004 03:29:44.266122 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fd1e584-06b3-4573-8ae5-cdcd9ed004e9" containerName="glance-log" Oct 04 03:29:44 crc kubenswrapper[4742]: I1004 03:29:44.266139 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fd1e584-06b3-4573-8ae5-cdcd9ed004e9" containerName="glance-httpd" Oct 04 03:29:44 crc kubenswrapper[4742]: I1004 03:29:44.266783 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-cm85w" Oct 04 03:29:44 crc kubenswrapper[4742]: I1004 03:29:44.320041 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-cm85w"] Oct 04 03:29:44 crc kubenswrapper[4742]: I1004 03:29:44.384529 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmx22\" (UniqueName: \"kubernetes.io/projected/8a438233-cd6d-4fbe-8008-d079ebd72d20-kube-api-access-lmx22\") pod \"glance-db-create-cm85w\" (UID: \"8a438233-cd6d-4fbe-8008-d079ebd72d20\") " pod="glance-kuttl-tests/glance-db-create-cm85w" Oct 04 03:29:44 crc kubenswrapper[4742]: I1004 03:29:44.487111 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmx22\" (UniqueName: \"kubernetes.io/projected/8a438233-cd6d-4fbe-8008-d079ebd72d20-kube-api-access-lmx22\") pod \"glance-db-create-cm85w\" (UID: \"8a438233-cd6d-4fbe-8008-d079ebd72d20\") " pod="glance-kuttl-tests/glance-db-create-cm85w" Oct 04 03:29:44 crc kubenswrapper[4742]: I1004 03:29:44.507474 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmx22\" (UniqueName: \"kubernetes.io/projected/8a438233-cd6d-4fbe-8008-d079ebd72d20-kube-api-access-lmx22\") pod \"glance-db-create-cm85w\" (UID: \"8a438233-cd6d-4fbe-8008-d079ebd72d20\") " pod="glance-kuttl-tests/glance-db-create-cm85w" Oct 04 03:29:44 crc kubenswrapper[4742]: I1004 03:29:44.589946 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-cm85w" Oct 04 03:29:44 crc kubenswrapper[4742]: I1004 03:29:44.829312 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-cm85w"] Oct 04 03:29:45 crc kubenswrapper[4742]: I1004 03:29:45.491534 4742 generic.go:334] "Generic (PLEG): container finished" podID="8a438233-cd6d-4fbe-8008-d079ebd72d20" containerID="daf788729c878af0db8fa608f6798c069e3449d29172b3b63ad756cb4e0a105c" exitCode=0 Oct 04 03:29:45 crc kubenswrapper[4742]: I1004 03:29:45.491737 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-cm85w" event={"ID":"8a438233-cd6d-4fbe-8008-d079ebd72d20","Type":"ContainerDied","Data":"daf788729c878af0db8fa608f6798c069e3449d29172b3b63ad756cb4e0a105c"} Oct 04 03:29:45 crc kubenswrapper[4742]: I1004 03:29:45.491980 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-cm85w" event={"ID":"8a438233-cd6d-4fbe-8008-d079ebd72d20","Type":"ContainerStarted","Data":"2270e8bd4dae1405ee6d7e3f85ef3e480824cb2c92bebc959e8bd0c9ba27a35e"} Oct 04 03:29:46 crc kubenswrapper[4742]: I1004 03:29:46.903223 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-cm85w" Oct 04 03:29:46 crc kubenswrapper[4742]: I1004 03:29:46.932243 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmx22\" (UniqueName: \"kubernetes.io/projected/8a438233-cd6d-4fbe-8008-d079ebd72d20-kube-api-access-lmx22\") pod \"8a438233-cd6d-4fbe-8008-d079ebd72d20\" (UID: \"8a438233-cd6d-4fbe-8008-d079ebd72d20\") " Oct 04 03:29:46 crc kubenswrapper[4742]: I1004 03:29:46.942638 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a438233-cd6d-4fbe-8008-d079ebd72d20-kube-api-access-lmx22" (OuterVolumeSpecName: "kube-api-access-lmx22") pod "8a438233-cd6d-4fbe-8008-d079ebd72d20" (UID: "8a438233-cd6d-4fbe-8008-d079ebd72d20"). InnerVolumeSpecName "kube-api-access-lmx22". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:29:47 crc kubenswrapper[4742]: I1004 03:29:47.034726 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmx22\" (UniqueName: \"kubernetes.io/projected/8a438233-cd6d-4fbe-8008-d079ebd72d20-kube-api-access-lmx22\") on node \"crc\" DevicePath \"\"" Oct 04 03:29:47 crc kubenswrapper[4742]: I1004 03:29:47.517257 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-cm85w" event={"ID":"8a438233-cd6d-4fbe-8008-d079ebd72d20","Type":"ContainerDied","Data":"2270e8bd4dae1405ee6d7e3f85ef3e480824cb2c92bebc959e8bd0c9ba27a35e"} Oct 04 03:29:47 crc kubenswrapper[4742]: I1004 03:29:47.517354 4742 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2270e8bd4dae1405ee6d7e3f85ef3e480824cb2c92bebc959e8bd0c9ba27a35e" Oct 04 03:29:47 crc kubenswrapper[4742]: I1004 03:29:47.517440 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-cm85w" Oct 04 03:29:54 crc kubenswrapper[4742]: I1004 03:29:54.288555 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-0305-account-create-frmkd"] Oct 04 03:29:54 crc kubenswrapper[4742]: E1004 03:29:54.289586 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a438233-cd6d-4fbe-8008-d079ebd72d20" containerName="mariadb-database-create" Oct 04 03:29:54 crc kubenswrapper[4742]: I1004 03:29:54.289607 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a438233-cd6d-4fbe-8008-d079ebd72d20" containerName="mariadb-database-create" Oct 04 03:29:54 crc kubenswrapper[4742]: I1004 03:29:54.289882 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a438233-cd6d-4fbe-8008-d079ebd72d20" containerName="mariadb-database-create" Oct 04 03:29:54 crc kubenswrapper[4742]: I1004 03:29:54.290560 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-0305-account-create-frmkd" Oct 04 03:29:54 crc kubenswrapper[4742]: I1004 03:29:54.292735 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Oct 04 03:29:54 crc kubenswrapper[4742]: I1004 03:29:54.313643 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-0305-account-create-frmkd"] Oct 04 03:29:54 crc kubenswrapper[4742]: I1004 03:29:54.401125 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m62cj\" (UniqueName: \"kubernetes.io/projected/fa270305-d910-44a7-956c-27b7268d448f-kube-api-access-m62cj\") pod \"glance-0305-account-create-frmkd\" (UID: \"fa270305-d910-44a7-956c-27b7268d448f\") " pod="glance-kuttl-tests/glance-0305-account-create-frmkd" Oct 04 03:29:54 crc kubenswrapper[4742]: I1004 03:29:54.504213 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m62cj\" (UniqueName: \"kubernetes.io/projected/fa270305-d910-44a7-956c-27b7268d448f-kube-api-access-m62cj\") pod \"glance-0305-account-create-frmkd\" (UID: \"fa270305-d910-44a7-956c-27b7268d448f\") " pod="glance-kuttl-tests/glance-0305-account-create-frmkd" Oct 04 03:29:54 crc kubenswrapper[4742]: I1004 03:29:54.528375 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m62cj\" (UniqueName: \"kubernetes.io/projected/fa270305-d910-44a7-956c-27b7268d448f-kube-api-access-m62cj\") pod \"glance-0305-account-create-frmkd\" (UID: \"fa270305-d910-44a7-956c-27b7268d448f\") " pod="glance-kuttl-tests/glance-0305-account-create-frmkd" Oct 04 03:29:54 crc kubenswrapper[4742]: I1004 03:29:54.620819 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-0305-account-create-frmkd" Oct 04 03:29:55 crc kubenswrapper[4742]: I1004 03:29:55.149678 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-0305-account-create-frmkd"] Oct 04 03:29:55 crc kubenswrapper[4742]: I1004 03:29:55.604801 4742 generic.go:334] "Generic (PLEG): container finished" podID="fa270305-d910-44a7-956c-27b7268d448f" containerID="8626a54ac197404e9c659c4d52753ead32c05a2d4db068d29eb3ad48882ee0fc" exitCode=0 Oct 04 03:29:55 crc kubenswrapper[4742]: I1004 03:29:55.604871 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-0305-account-create-frmkd" event={"ID":"fa270305-d910-44a7-956c-27b7268d448f","Type":"ContainerDied","Data":"8626a54ac197404e9c659c4d52753ead32c05a2d4db068d29eb3ad48882ee0fc"} Oct 04 03:29:55 crc kubenswrapper[4742]: I1004 03:29:55.604912 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-0305-account-create-frmkd" event={"ID":"fa270305-d910-44a7-956c-27b7268d448f","Type":"ContainerStarted","Data":"cba5a071879eee402abf0988b3599ee89add6b58fad37e45042d2f0c6e4b8b51"} Oct 04 03:29:57 crc kubenswrapper[4742]: I1004 03:29:57.029194 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-0305-account-create-frmkd" Oct 04 03:29:57 crc kubenswrapper[4742]: I1004 03:29:57.148614 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m62cj\" (UniqueName: \"kubernetes.io/projected/fa270305-d910-44a7-956c-27b7268d448f-kube-api-access-m62cj\") pod \"fa270305-d910-44a7-956c-27b7268d448f\" (UID: \"fa270305-d910-44a7-956c-27b7268d448f\") " Oct 04 03:29:57 crc kubenswrapper[4742]: I1004 03:29:57.156529 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa270305-d910-44a7-956c-27b7268d448f-kube-api-access-m62cj" (OuterVolumeSpecName: "kube-api-access-m62cj") pod "fa270305-d910-44a7-956c-27b7268d448f" (UID: "fa270305-d910-44a7-956c-27b7268d448f"). InnerVolumeSpecName "kube-api-access-m62cj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:29:57 crc kubenswrapper[4742]: I1004 03:29:57.251761 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m62cj\" (UniqueName: \"kubernetes.io/projected/fa270305-d910-44a7-956c-27b7268d448f-kube-api-access-m62cj\") on node \"crc\" DevicePath \"\"" Oct 04 03:29:57 crc kubenswrapper[4742]: I1004 03:29:57.627124 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-0305-account-create-frmkd" event={"ID":"fa270305-d910-44a7-956c-27b7268d448f","Type":"ContainerDied","Data":"cba5a071879eee402abf0988b3599ee89add6b58fad37e45042d2f0c6e4b8b51"} Oct 04 03:29:57 crc kubenswrapper[4742]: I1004 03:29:57.627189 4742 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cba5a071879eee402abf0988b3599ee89add6b58fad37e45042d2f0c6e4b8b51" Oct 04 03:29:57 crc kubenswrapper[4742]: I1004 03:29:57.627190 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-0305-account-create-frmkd" Oct 04 03:29:59 crc kubenswrapper[4742]: I1004 03:29:59.441968 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-mwk7z"] Oct 04 03:29:59 crc kubenswrapper[4742]: E1004 03:29:59.442316 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa270305-d910-44a7-956c-27b7268d448f" containerName="mariadb-account-create" Oct 04 03:29:59 crc kubenswrapper[4742]: I1004 03:29:59.442332 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa270305-d910-44a7-956c-27b7268d448f" containerName="mariadb-account-create" Oct 04 03:29:59 crc kubenswrapper[4742]: I1004 03:29:59.442539 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa270305-d910-44a7-956c-27b7268d448f" containerName="mariadb-account-create" Oct 04 03:29:59 crc kubenswrapper[4742]: I1004 03:29:59.443114 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-mwk7z" Oct 04 03:29:59 crc kubenswrapper[4742]: I1004 03:29:59.453957 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-8fjsr" Oct 04 03:29:59 crc kubenswrapper[4742]: I1004 03:29:59.454172 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Oct 04 03:29:59 crc kubenswrapper[4742]: I1004 03:29:59.460666 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-mwk7z"] Oct 04 03:29:59 crc kubenswrapper[4742]: I1004 03:29:59.488141 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjhjv\" (UniqueName: \"kubernetes.io/projected/679af79d-4438-49e7-ace8-d83ff23a2a97-kube-api-access-bjhjv\") pod \"glance-db-sync-mwk7z\" (UID: \"679af79d-4438-49e7-ace8-d83ff23a2a97\") " pod="glance-kuttl-tests/glance-db-sync-mwk7z" Oct 04 03:29:59 crc kubenswrapper[4742]: I1004 03:29:59.488394 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/679af79d-4438-49e7-ace8-d83ff23a2a97-config-data\") pod \"glance-db-sync-mwk7z\" (UID: \"679af79d-4438-49e7-ace8-d83ff23a2a97\") " pod="glance-kuttl-tests/glance-db-sync-mwk7z" Oct 04 03:29:59 crc kubenswrapper[4742]: I1004 03:29:59.488509 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/679af79d-4438-49e7-ace8-d83ff23a2a97-db-sync-config-data\") pod \"glance-db-sync-mwk7z\" (UID: \"679af79d-4438-49e7-ace8-d83ff23a2a97\") " pod="glance-kuttl-tests/glance-db-sync-mwk7z" Oct 04 03:29:59 crc kubenswrapper[4742]: I1004 03:29:59.589641 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/679af79d-4438-49e7-ace8-d83ff23a2a97-db-sync-config-data\") pod \"glance-db-sync-mwk7z\" (UID: \"679af79d-4438-49e7-ace8-d83ff23a2a97\") " pod="glance-kuttl-tests/glance-db-sync-mwk7z" Oct 04 03:29:59 crc kubenswrapper[4742]: I1004 03:29:59.589917 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjhjv\" (UniqueName: \"kubernetes.io/projected/679af79d-4438-49e7-ace8-d83ff23a2a97-kube-api-access-bjhjv\") pod \"glance-db-sync-mwk7z\" (UID: \"679af79d-4438-49e7-ace8-d83ff23a2a97\") " pod="glance-kuttl-tests/glance-db-sync-mwk7z" Oct 04 03:29:59 crc kubenswrapper[4742]: I1004 03:29:59.590078 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/679af79d-4438-49e7-ace8-d83ff23a2a97-config-data\") pod \"glance-db-sync-mwk7z\" (UID: \"679af79d-4438-49e7-ace8-d83ff23a2a97\") " pod="glance-kuttl-tests/glance-db-sync-mwk7z" Oct 04 03:29:59 crc kubenswrapper[4742]: I1004 03:29:59.595136 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/679af79d-4438-49e7-ace8-d83ff23a2a97-config-data\") pod \"glance-db-sync-mwk7z\" (UID: \"679af79d-4438-49e7-ace8-d83ff23a2a97\") " pod="glance-kuttl-tests/glance-db-sync-mwk7z" Oct 04 03:29:59 crc kubenswrapper[4742]: I1004 03:29:59.610092 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/679af79d-4438-49e7-ace8-d83ff23a2a97-db-sync-config-data\") pod \"glance-db-sync-mwk7z\" (UID: \"679af79d-4438-49e7-ace8-d83ff23a2a97\") " pod="glance-kuttl-tests/glance-db-sync-mwk7z" Oct 04 03:29:59 crc kubenswrapper[4742]: I1004 03:29:59.617028 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjhjv\" (UniqueName: \"kubernetes.io/projected/679af79d-4438-49e7-ace8-d83ff23a2a97-kube-api-access-bjhjv\") pod \"glance-db-sync-mwk7z\" (UID: \"679af79d-4438-49e7-ace8-d83ff23a2a97\") " pod="glance-kuttl-tests/glance-db-sync-mwk7z" Oct 04 03:29:59 crc kubenswrapper[4742]: I1004 03:29:59.771038 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-mwk7z" Oct 04 03:30:00 crc kubenswrapper[4742]: I1004 03:30:00.149618 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325810-4qftb"] Oct 04 03:30:00 crc kubenswrapper[4742]: I1004 03:30:00.151225 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-4qftb" Oct 04 03:30:00 crc kubenswrapper[4742]: I1004 03:30:00.153471 4742 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 03:30:00 crc kubenswrapper[4742]: I1004 03:30:00.160978 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 03:30:00 crc kubenswrapper[4742]: I1004 03:30:00.161729 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325810-4qftb"] Oct 04 03:30:00 crc kubenswrapper[4742]: I1004 03:30:00.203674 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f4732f31-b888-4a74-9cc5-bf196456859a-config-volume\") pod \"collect-profiles-29325810-4qftb\" (UID: \"f4732f31-b888-4a74-9cc5-bf196456859a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-4qftb" Oct 04 03:30:00 crc kubenswrapper[4742]: I1004 03:30:00.203754 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgcpg\" (UniqueName: \"kubernetes.io/projected/f4732f31-b888-4a74-9cc5-bf196456859a-kube-api-access-bgcpg\") pod \"collect-profiles-29325810-4qftb\" (UID: \"f4732f31-b888-4a74-9cc5-bf196456859a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-4qftb" Oct 04 03:30:00 crc kubenswrapper[4742]: I1004 03:30:00.203806 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f4732f31-b888-4a74-9cc5-bf196456859a-secret-volume\") pod \"collect-profiles-29325810-4qftb\" (UID: \"f4732f31-b888-4a74-9cc5-bf196456859a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-4qftb" Oct 04 03:30:00 crc kubenswrapper[4742]: I1004 03:30:00.269777 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-mwk7z"] Oct 04 03:30:00 crc kubenswrapper[4742]: I1004 03:30:00.305225 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f4732f31-b888-4a74-9cc5-bf196456859a-config-volume\") pod \"collect-profiles-29325810-4qftb\" (UID: \"f4732f31-b888-4a74-9cc5-bf196456859a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-4qftb" Oct 04 03:30:00 crc kubenswrapper[4742]: I1004 03:30:00.305318 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgcpg\" (UniqueName: \"kubernetes.io/projected/f4732f31-b888-4a74-9cc5-bf196456859a-kube-api-access-bgcpg\") pod \"collect-profiles-29325810-4qftb\" (UID: \"f4732f31-b888-4a74-9cc5-bf196456859a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-4qftb" Oct 04 03:30:00 crc kubenswrapper[4742]: I1004 03:30:00.305373 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f4732f31-b888-4a74-9cc5-bf196456859a-secret-volume\") pod \"collect-profiles-29325810-4qftb\" (UID: \"f4732f31-b888-4a74-9cc5-bf196456859a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-4qftb" Oct 04 03:30:00 crc kubenswrapper[4742]: I1004 03:30:00.307094 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f4732f31-b888-4a74-9cc5-bf196456859a-config-volume\") pod \"collect-profiles-29325810-4qftb\" (UID: \"f4732f31-b888-4a74-9cc5-bf196456859a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-4qftb" Oct 04 03:30:00 crc kubenswrapper[4742]: I1004 03:30:00.309555 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f4732f31-b888-4a74-9cc5-bf196456859a-secret-volume\") pod \"collect-profiles-29325810-4qftb\" (UID: \"f4732f31-b888-4a74-9cc5-bf196456859a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-4qftb" Oct 04 03:30:00 crc kubenswrapper[4742]: I1004 03:30:00.326564 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgcpg\" (UniqueName: \"kubernetes.io/projected/f4732f31-b888-4a74-9cc5-bf196456859a-kube-api-access-bgcpg\") pod \"collect-profiles-29325810-4qftb\" (UID: \"f4732f31-b888-4a74-9cc5-bf196456859a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-4qftb" Oct 04 03:30:00 crc kubenswrapper[4742]: I1004 03:30:00.485127 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-4qftb" Oct 04 03:30:00 crc kubenswrapper[4742]: I1004 03:30:00.653079 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-mwk7z" event={"ID":"679af79d-4438-49e7-ace8-d83ff23a2a97","Type":"ContainerStarted","Data":"e089fa6fd165a87c8713700abfa8045673a7acf90ff88b4ae37beeb48352da7b"} Oct 04 03:30:00 crc kubenswrapper[4742]: I1004 03:30:00.946376 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325810-4qftb"] Oct 04 03:30:01 crc kubenswrapper[4742]: I1004 03:30:01.663970 4742 generic.go:334] "Generic (PLEG): container finished" podID="f4732f31-b888-4a74-9cc5-bf196456859a" containerID="7247623cc77be439b88c170373d90c39195e8d53fe23921a4cf1a1278878bb5d" exitCode=0 Oct 04 03:30:01 crc kubenswrapper[4742]: I1004 03:30:01.664054 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-4qftb" event={"ID":"f4732f31-b888-4a74-9cc5-bf196456859a","Type":"ContainerDied","Data":"7247623cc77be439b88c170373d90c39195e8d53fe23921a4cf1a1278878bb5d"} Oct 04 03:30:01 crc kubenswrapper[4742]: I1004 03:30:01.664100 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-4qftb" event={"ID":"f4732f31-b888-4a74-9cc5-bf196456859a","Type":"ContainerStarted","Data":"d82d9aa757ac12b9f23c35e8c086731920a05dc8c630a22ff08e0e55488e6613"} Oct 04 03:30:01 crc kubenswrapper[4742]: I1004 03:30:01.665760 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-mwk7z" event={"ID":"679af79d-4438-49e7-ace8-d83ff23a2a97","Type":"ContainerStarted","Data":"6e451cad86f2dfcf4fdafd7eda8e2daf1c97f2a8ebb5e09649553cdfc7c2ac3e"} Oct 04 03:30:01 crc kubenswrapper[4742]: I1004 03:30:01.701496 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-mwk7z" podStartSLOduration=2.70147226 podStartE2EDuration="2.70147226s" podCreationTimestamp="2025-10-04 03:29:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:30:01.696433031 +0000 UTC m=+1167.628770207" watchObservedRunningTime="2025-10-04 03:30:01.70147226 +0000 UTC m=+1167.633809436" Oct 04 03:30:02 crc kubenswrapper[4742]: I1004 03:30:02.987763 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-4qftb" Oct 04 03:30:03 crc kubenswrapper[4742]: I1004 03:30:03.047861 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgcpg\" (UniqueName: \"kubernetes.io/projected/f4732f31-b888-4a74-9cc5-bf196456859a-kube-api-access-bgcpg\") pod \"f4732f31-b888-4a74-9cc5-bf196456859a\" (UID: \"f4732f31-b888-4a74-9cc5-bf196456859a\") " Oct 04 03:30:03 crc kubenswrapper[4742]: I1004 03:30:03.047940 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f4732f31-b888-4a74-9cc5-bf196456859a-secret-volume\") pod \"f4732f31-b888-4a74-9cc5-bf196456859a\" (UID: \"f4732f31-b888-4a74-9cc5-bf196456859a\") " Oct 04 03:30:03 crc kubenswrapper[4742]: I1004 03:30:03.047980 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f4732f31-b888-4a74-9cc5-bf196456859a-config-volume\") pod \"f4732f31-b888-4a74-9cc5-bf196456859a\" (UID: \"f4732f31-b888-4a74-9cc5-bf196456859a\") " Oct 04 03:30:03 crc kubenswrapper[4742]: I1004 03:30:03.049131 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4732f31-b888-4a74-9cc5-bf196456859a-config-volume" (OuterVolumeSpecName: "config-volume") pod "f4732f31-b888-4a74-9cc5-bf196456859a" (UID: "f4732f31-b888-4a74-9cc5-bf196456859a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:30:03 crc kubenswrapper[4742]: I1004 03:30:03.056565 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4732f31-b888-4a74-9cc5-bf196456859a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f4732f31-b888-4a74-9cc5-bf196456859a" (UID: "f4732f31-b888-4a74-9cc5-bf196456859a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:30:03 crc kubenswrapper[4742]: I1004 03:30:03.057052 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4732f31-b888-4a74-9cc5-bf196456859a-kube-api-access-bgcpg" (OuterVolumeSpecName: "kube-api-access-bgcpg") pod "f4732f31-b888-4a74-9cc5-bf196456859a" (UID: "f4732f31-b888-4a74-9cc5-bf196456859a"). InnerVolumeSpecName "kube-api-access-bgcpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:30:03 crc kubenswrapper[4742]: I1004 03:30:03.149776 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgcpg\" (UniqueName: \"kubernetes.io/projected/f4732f31-b888-4a74-9cc5-bf196456859a-kube-api-access-bgcpg\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:03 crc kubenswrapper[4742]: I1004 03:30:03.149828 4742 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f4732f31-b888-4a74-9cc5-bf196456859a-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:03 crc kubenswrapper[4742]: I1004 03:30:03.149840 4742 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f4732f31-b888-4a74-9cc5-bf196456859a-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:03 crc kubenswrapper[4742]: I1004 03:30:03.683375 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-4qftb" event={"ID":"f4732f31-b888-4a74-9cc5-bf196456859a","Type":"ContainerDied","Data":"d82d9aa757ac12b9f23c35e8c086731920a05dc8c630a22ff08e0e55488e6613"} Oct 04 03:30:03 crc kubenswrapper[4742]: I1004 03:30:03.684078 4742 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d82d9aa757ac12b9f23c35e8c086731920a05dc8c630a22ff08e0e55488e6613" Oct 04 03:30:03 crc kubenswrapper[4742]: I1004 03:30:03.683490 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-4qftb" Oct 04 03:30:04 crc kubenswrapper[4742]: I1004 03:30:04.707256 4742 generic.go:334] "Generic (PLEG): container finished" podID="679af79d-4438-49e7-ace8-d83ff23a2a97" containerID="6e451cad86f2dfcf4fdafd7eda8e2daf1c97f2a8ebb5e09649553cdfc7c2ac3e" exitCode=0 Oct 04 03:30:04 crc kubenswrapper[4742]: I1004 03:30:04.707410 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-mwk7z" event={"ID":"679af79d-4438-49e7-ace8-d83ff23a2a97","Type":"ContainerDied","Data":"6e451cad86f2dfcf4fdafd7eda8e2daf1c97f2a8ebb5e09649553cdfc7c2ac3e"} Oct 04 03:30:06 crc kubenswrapper[4742]: I1004 03:30:06.153913 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-mwk7z" Oct 04 03:30:06 crc kubenswrapper[4742]: I1004 03:30:06.205025 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjhjv\" (UniqueName: \"kubernetes.io/projected/679af79d-4438-49e7-ace8-d83ff23a2a97-kube-api-access-bjhjv\") pod \"679af79d-4438-49e7-ace8-d83ff23a2a97\" (UID: \"679af79d-4438-49e7-ace8-d83ff23a2a97\") " Oct 04 03:30:06 crc kubenswrapper[4742]: I1004 03:30:06.205220 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/679af79d-4438-49e7-ace8-d83ff23a2a97-db-sync-config-data\") pod \"679af79d-4438-49e7-ace8-d83ff23a2a97\" (UID: \"679af79d-4438-49e7-ace8-d83ff23a2a97\") " Oct 04 03:30:06 crc kubenswrapper[4742]: I1004 03:30:06.205301 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/679af79d-4438-49e7-ace8-d83ff23a2a97-config-data\") pod \"679af79d-4438-49e7-ace8-d83ff23a2a97\" (UID: \"679af79d-4438-49e7-ace8-d83ff23a2a97\") " Oct 04 03:30:06 crc kubenswrapper[4742]: I1004 03:30:06.218193 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/679af79d-4438-49e7-ace8-d83ff23a2a97-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "679af79d-4438-49e7-ace8-d83ff23a2a97" (UID: "679af79d-4438-49e7-ace8-d83ff23a2a97"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:30:06 crc kubenswrapper[4742]: I1004 03:30:06.220074 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/679af79d-4438-49e7-ace8-d83ff23a2a97-kube-api-access-bjhjv" (OuterVolumeSpecName: "kube-api-access-bjhjv") pod "679af79d-4438-49e7-ace8-d83ff23a2a97" (UID: "679af79d-4438-49e7-ace8-d83ff23a2a97"). InnerVolumeSpecName "kube-api-access-bjhjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:30:06 crc kubenswrapper[4742]: I1004 03:30:06.282237 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/679af79d-4438-49e7-ace8-d83ff23a2a97-config-data" (OuterVolumeSpecName: "config-data") pod "679af79d-4438-49e7-ace8-d83ff23a2a97" (UID: "679af79d-4438-49e7-ace8-d83ff23a2a97"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:30:06 crc kubenswrapper[4742]: I1004 03:30:06.308413 4742 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/679af79d-4438-49e7-ace8-d83ff23a2a97-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:06 crc kubenswrapper[4742]: I1004 03:30:06.308476 4742 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/679af79d-4438-49e7-ace8-d83ff23a2a97-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:06 crc kubenswrapper[4742]: I1004 03:30:06.308497 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjhjv\" (UniqueName: \"kubernetes.io/projected/679af79d-4438-49e7-ace8-d83ff23a2a97-kube-api-access-bjhjv\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:06 crc kubenswrapper[4742]: I1004 03:30:06.736022 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-mwk7z" event={"ID":"679af79d-4438-49e7-ace8-d83ff23a2a97","Type":"ContainerDied","Data":"e089fa6fd165a87c8713700abfa8045673a7acf90ff88b4ae37beeb48352da7b"} Oct 04 03:30:06 crc kubenswrapper[4742]: I1004 03:30:06.736096 4742 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e089fa6fd165a87c8713700abfa8045673a7acf90ff88b4ae37beeb48352da7b" Oct 04 03:30:06 crc kubenswrapper[4742]: I1004 03:30:06.736152 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-mwk7z" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.157529 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Oct 04 03:30:08 crc kubenswrapper[4742]: E1004 03:30:08.157933 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="679af79d-4438-49e7-ace8-d83ff23a2a97" containerName="glance-db-sync" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.157950 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="679af79d-4438-49e7-ace8-d83ff23a2a97" containerName="glance-db-sync" Oct 04 03:30:08 crc kubenswrapper[4742]: E1004 03:30:08.157990 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4732f31-b888-4a74-9cc5-bf196456859a" containerName="collect-profiles" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.157996 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4732f31-b888-4a74-9cc5-bf196456859a" containerName="collect-profiles" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.158140 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="679af79d-4438-49e7-ace8-d83ff23a2a97" containerName="glance-db-sync" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.158159 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4732f31-b888-4a74-9cc5-bf196456859a" containerName="collect-profiles" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.164862 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.168474 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.170465 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-8fjsr" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.171243 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-external-config-data" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.199736 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.217550 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.218773 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.222615 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.243921 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.268265 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.283174 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Oct 04 03:30:08 crc kubenswrapper[4742]: E1004 03:30:08.283685 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config-data dev etc-iscsi etc-nvme glance glance-cache httpd-run kube-api-access-n59gl lib-modules logs run scripts sys var-locks-brick], unattached volumes=[], failed to process volumes=[]: context canceled" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="3e4971fb-098a-4711-a4f6-370b1a5a1120" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.347790 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-run\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.347878 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-dev\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.347917 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.347955 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.348004 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.348030 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3e4971fb-098a-4711-a4f6-370b1a5a1120-logs\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.348049 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.348077 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.348119 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e4971fb-098a-4711-a4f6-370b1a5a1120-config-data\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.348148 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e4971fb-098a-4711-a4f6-370b1a5a1120-scripts\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.348170 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n59gl\" (UniqueName: \"kubernetes.io/projected/3e4971fb-098a-4711-a4f6-370b1a5a1120-kube-api-access-n59gl\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.348198 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbzcw\" (UniqueName: \"kubernetes.io/projected/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-kube-api-access-xbzcw\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.348224 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-dev\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.348247 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.348946 4742 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") device mount path \"/mnt/openstack/pv11\"" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.348267 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-run\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.367570 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.367657 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.367709 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.367751 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-logs\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.367889 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.367931 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.367965 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.368019 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-sys\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.368063 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.368132 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-sys\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.368236 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.368268 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3e4971fb-098a-4711-a4f6-370b1a5a1120-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.368317 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.403146 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.469806 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-dev\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.470177 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.470197 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-run\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.470219 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.470236 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.470257 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.470322 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-logs\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.470348 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.470374 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.470403 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.470425 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-sys\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.470444 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.470473 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-sys\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.470501 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.470521 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3e4971fb-098a-4711-a4f6-370b1a5a1120-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.470544 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.470567 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-run\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.470588 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-dev\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.470611 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.470639 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.471517 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3e4971fb-098a-4711-a4f6-370b1a5a1120-logs\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.470791 4742 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.471549 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.471573 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.471592 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e4971fb-098a-4711-a4f6-370b1a5a1120-config-data\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.471620 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e4971fb-098a-4711-a4f6-370b1a5a1120-scripts\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.471647 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n59gl\" (UniqueName: \"kubernetes.io/projected/3e4971fb-098a-4711-a4f6-370b1a5a1120-kube-api-access-n59gl\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.471669 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbzcw\" (UniqueName: \"kubernetes.io/projected/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-kube-api-access-xbzcw\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.472367 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3e4971fb-098a-4711-a4f6-370b1a5a1120-logs\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.470817 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-dev\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.472435 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.470867 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-run\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.472478 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.470886 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.474940 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-sys\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.475130 4742 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.475152 4742 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") device mount path \"/mnt/openstack/pv12\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.476589 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.476662 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.476888 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3e4971fb-098a-4711-a4f6-370b1a5a1120-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.477576 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-dev\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.477620 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-run\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.477888 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.478065 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e4971fb-098a-4711-a4f6-370b1a5a1120-config-data\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.478108 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-logs\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.478115 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-sys\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.478149 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.470904 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.478152 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.470848 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.482232 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e4971fb-098a-4711-a4f6-370b1a5a1120-scripts\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.485073 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.492863 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbzcw\" (UniqueName: \"kubernetes.io/projected/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-kube-api-access-xbzcw\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.495730 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n59gl\" (UniqueName: \"kubernetes.io/projected/3e4971fb-098a-4711-a4f6-370b1a5a1120-kube-api-access-n59gl\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.503797 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.506917 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.507385 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.538975 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.755385 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.771485 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.777172 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e4971fb-098a-4711-a4f6-370b1a5a1120-config-data\") pod \"3e4971fb-098a-4711-a4f6-370b1a5a1120\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.777292 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n59gl\" (UniqueName: \"kubernetes.io/projected/3e4971fb-098a-4711-a4f6-370b1a5a1120-kube-api-access-n59gl\") pod \"3e4971fb-098a-4711-a4f6-370b1a5a1120\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.777336 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"3e4971fb-098a-4711-a4f6-370b1a5a1120\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.777357 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-dev\") pod \"3e4971fb-098a-4711-a4f6-370b1a5a1120\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.777379 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e4971fb-098a-4711-a4f6-370b1a5a1120-scripts\") pod \"3e4971fb-098a-4711-a4f6-370b1a5a1120\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.777413 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"3e4971fb-098a-4711-a4f6-370b1a5a1120\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.777481 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3e4971fb-098a-4711-a4f6-370b1a5a1120-logs\") pod \"3e4971fb-098a-4711-a4f6-370b1a5a1120\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.777479 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-dev" (OuterVolumeSpecName: "dev") pod "3e4971fb-098a-4711-a4f6-370b1a5a1120" (UID: "3e4971fb-098a-4711-a4f6-370b1a5a1120"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.777508 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-etc-iscsi\") pod \"3e4971fb-098a-4711-a4f6-370b1a5a1120\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.777575 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-lib-modules\") pod \"3e4971fb-098a-4711-a4f6-370b1a5a1120\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.777802 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-var-locks-brick\") pod \"3e4971fb-098a-4711-a4f6-370b1a5a1120\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.777830 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-sys\") pod \"3e4971fb-098a-4711-a4f6-370b1a5a1120\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.777940 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-etc-nvme\") pod \"3e4971fb-098a-4711-a4f6-370b1a5a1120\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.778000 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3e4971fb-098a-4711-a4f6-370b1a5a1120-httpd-run\") pod \"3e4971fb-098a-4711-a4f6-370b1a5a1120\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.778038 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-run\") pod \"3e4971fb-098a-4711-a4f6-370b1a5a1120\" (UID: \"3e4971fb-098a-4711-a4f6-370b1a5a1120\") " Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.778057 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "3e4971fb-098a-4711-a4f6-370b1a5a1120" (UID: "3e4971fb-098a-4711-a4f6-370b1a5a1120"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.778398 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e4971fb-098a-4711-a4f6-370b1a5a1120-logs" (OuterVolumeSpecName: "logs") pod "3e4971fb-098a-4711-a4f6-370b1a5a1120" (UID: "3e4971fb-098a-4711-a4f6-370b1a5a1120"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.778433 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "3e4971fb-098a-4711-a4f6-370b1a5a1120" (UID: "3e4971fb-098a-4711-a4f6-370b1a5a1120"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.778452 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "3e4971fb-098a-4711-a4f6-370b1a5a1120" (UID: "3e4971fb-098a-4711-a4f6-370b1a5a1120"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.778465 4742 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-var-locks-brick\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.778483 4742 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-dev\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.778482 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "3e4971fb-098a-4711-a4f6-370b1a5a1120" (UID: "3e4971fb-098a-4711-a4f6-370b1a5a1120"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.778496 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-sys" (OuterVolumeSpecName: "sys") pod "3e4971fb-098a-4711-a4f6-370b1a5a1120" (UID: "3e4971fb-098a-4711-a4f6-370b1a5a1120"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.778525 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-run" (OuterVolumeSpecName: "run") pod "3e4971fb-098a-4711-a4f6-370b1a5a1120" (UID: "3e4971fb-098a-4711-a4f6-370b1a5a1120"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.778706 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e4971fb-098a-4711-a4f6-370b1a5a1120-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3e4971fb-098a-4711-a4f6-370b1a5a1120" (UID: "3e4971fb-098a-4711-a4f6-370b1a5a1120"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.781708 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "3e4971fb-098a-4711-a4f6-370b1a5a1120" (UID: "3e4971fb-098a-4711-a4f6-370b1a5a1120"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.781738 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance-cache") pod "3e4971fb-098a-4711-a4f6-370b1a5a1120" (UID: "3e4971fb-098a-4711-a4f6-370b1a5a1120"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.782571 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e4971fb-098a-4711-a4f6-370b1a5a1120-scripts" (OuterVolumeSpecName: "scripts") pod "3e4971fb-098a-4711-a4f6-370b1a5a1120" (UID: "3e4971fb-098a-4711-a4f6-370b1a5a1120"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.782909 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e4971fb-098a-4711-a4f6-370b1a5a1120-kube-api-access-n59gl" (OuterVolumeSpecName: "kube-api-access-n59gl") pod "3e4971fb-098a-4711-a4f6-370b1a5a1120" (UID: "3e4971fb-098a-4711-a4f6-370b1a5a1120"). InnerVolumeSpecName "kube-api-access-n59gl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.783468 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e4971fb-098a-4711-a4f6-370b1a5a1120-config-data" (OuterVolumeSpecName: "config-data") pod "3e4971fb-098a-4711-a4f6-370b1a5a1120" (UID: "3e4971fb-098a-4711-a4f6-370b1a5a1120"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.880340 4742 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-etc-nvme\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.880376 4742 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3e4971fb-098a-4711-a4f6-370b1a5a1120-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.880410 4742 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-run\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.880421 4742 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e4971fb-098a-4711-a4f6-370b1a5a1120-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.880432 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n59gl\" (UniqueName: \"kubernetes.io/projected/3e4971fb-098a-4711-a4f6-370b1a5a1120-kube-api-access-n59gl\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.880456 4742 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.880486 4742 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e4971fb-098a-4711-a4f6-370b1a5a1120-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.880502 4742 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.880512 4742 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3e4971fb-098a-4711-a4f6-370b1a5a1120-logs\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.880523 4742 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-etc-iscsi\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.880532 4742 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-lib-modules\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.880560 4742 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3e4971fb-098a-4711-a4f6-370b1a5a1120-sys\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.897195 4742 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.905960 4742 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.981662 4742 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:08 crc kubenswrapper[4742]: I1004 03:30:08.981700 4742 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:09 crc kubenswrapper[4742]: I1004 03:30:09.019363 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Oct 04 03:30:09 crc kubenswrapper[4742]: I1004 03:30:09.769005 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:09 crc kubenswrapper[4742]: I1004 03:30:09.769537 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489","Type":"ContainerStarted","Data":"ff221474def26197448aede829b1e388fc8ddf74d90f7eb9d6bcaf18e436fe70"} Oct 04 03:30:09 crc kubenswrapper[4742]: I1004 03:30:09.769952 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489","Type":"ContainerStarted","Data":"cfd40cf0b885b8df11a6da79255f67c0d823f46256e9556716c960b0e8e81ad4"} Oct 04 03:30:09 crc kubenswrapper[4742]: I1004 03:30:09.769970 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489","Type":"ContainerStarted","Data":"fa19eaa8e6a0fba7017dde3d78e5ae7240f4ccac449bca91b170f44eb00f8ce1"} Oct 04 03:30:09 crc kubenswrapper[4742]: I1004 03:30:09.769980 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489","Type":"ContainerStarted","Data":"c178df44d8da0b563563360cc55209fcee77fdecfcab82e34b5e8475bb36f7a8"} Oct 04 03:30:09 crc kubenswrapper[4742]: I1004 03:30:09.804597 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=2.804576837 podStartE2EDuration="2.804576837s" podCreationTimestamp="2025-10-04 03:30:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:30:09.800162713 +0000 UTC m=+1175.732499889" watchObservedRunningTime="2025-10-04 03:30:09.804576837 +0000 UTC m=+1175.736914013" Oct 04 03:30:09 crc kubenswrapper[4742]: I1004 03:30:09.856823 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Oct 04 03:30:09 crc kubenswrapper[4742]: I1004 03:30:09.863862 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Oct 04 03:30:09 crc kubenswrapper[4742]: I1004 03:30:09.896506 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Oct 04 03:30:09 crc kubenswrapper[4742]: I1004 03:30:09.897905 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:09 crc kubenswrapper[4742]: I1004 03:30:09.913313 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-external-config-data" Oct 04 03:30:09 crc kubenswrapper[4742]: I1004 03:30:09.923971 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Oct 04 03:30:09 crc kubenswrapper[4742]: I1004 03:30:09.932850 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.002317 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqgrr\" (UniqueName: \"kubernetes.io/projected/7cad2cc9-60ac-4e19-81cc-4676387a2176-kube-api-access-cqgrr\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.002702 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cad2cc9-60ac-4e19-81cc-4676387a2176-logs\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.002787 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7cad2cc9-60ac-4e19-81cc-4676387a2176-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.002879 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.002961 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7cad2cc9-60ac-4e19-81cc-4676387a2176-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.003060 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7cad2cc9-60ac-4e19-81cc-4676387a2176-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.003161 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cad2cc9-60ac-4e19-81cc-4676387a2176-scripts\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.003238 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7cad2cc9-60ac-4e19-81cc-4676387a2176-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.003370 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7cad2cc9-60ac-4e19-81cc-4676387a2176-sys\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.003496 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7cad2cc9-60ac-4e19-81cc-4676387a2176-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.003564 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.004779 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cad2cc9-60ac-4e19-81cc-4676387a2176-config-data\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.004829 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7cad2cc9-60ac-4e19-81cc-4676387a2176-run\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.004936 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7cad2cc9-60ac-4e19-81cc-4676387a2176-dev\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.106778 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7cad2cc9-60ac-4e19-81cc-4676387a2176-sys\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.106853 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7cad2cc9-60ac-4e19-81cc-4676387a2176-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.106892 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.106942 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cad2cc9-60ac-4e19-81cc-4676387a2176-config-data\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.106973 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7cad2cc9-60ac-4e19-81cc-4676387a2176-run\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.106994 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7cad2cc9-60ac-4e19-81cc-4676387a2176-dev\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.107065 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqgrr\" (UniqueName: \"kubernetes.io/projected/7cad2cc9-60ac-4e19-81cc-4676387a2176-kube-api-access-cqgrr\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.107091 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cad2cc9-60ac-4e19-81cc-4676387a2176-logs\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.107081 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7cad2cc9-60ac-4e19-81cc-4676387a2176-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.107444 4742 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") device mount path \"/mnt/openstack/pv11\"" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.106935 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7cad2cc9-60ac-4e19-81cc-4676387a2176-sys\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.107812 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7cad2cc9-60ac-4e19-81cc-4676387a2176-run\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.107835 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7cad2cc9-60ac-4e19-81cc-4676387a2176-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.107115 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7cad2cc9-60ac-4e19-81cc-4676387a2176-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.107937 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.107986 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7cad2cc9-60ac-4e19-81cc-4676387a2176-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.108094 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7cad2cc9-60ac-4e19-81cc-4676387a2176-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.108115 4742 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.108146 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7cad2cc9-60ac-4e19-81cc-4676387a2176-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.108219 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7cad2cc9-60ac-4e19-81cc-4676387a2176-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.108287 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cad2cc9-60ac-4e19-81cc-4676387a2176-scripts\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.108310 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7cad2cc9-60ac-4e19-81cc-4676387a2176-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.108375 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cad2cc9-60ac-4e19-81cc-4676387a2176-logs\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.108388 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7cad2cc9-60ac-4e19-81cc-4676387a2176-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.108101 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7cad2cc9-60ac-4e19-81cc-4676387a2176-dev\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.117832 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cad2cc9-60ac-4e19-81cc-4676387a2176-config-data\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.117902 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cad2cc9-60ac-4e19-81cc-4676387a2176-scripts\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.126674 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqgrr\" (UniqueName: \"kubernetes.io/projected/7cad2cc9-60ac-4e19-81cc-4676387a2176-kube-api-access-cqgrr\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.138054 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.151075 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"7cad2cc9-60ac-4e19-81cc-4676387a2176\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.227459 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.540782 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.792251 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"7cad2cc9-60ac-4e19-81cc-4676387a2176","Type":"ContainerStarted","Data":"c8e5788bd2c333b96c875f6c0af676dbe87e1736d8569db58cefaf9ebdeecb21"} Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.793085 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"7cad2cc9-60ac-4e19-81cc-4676387a2176","Type":"ContainerStarted","Data":"aaab4ed213aa516ef2d0c2866a62829fc2cbdddc525ade83d145626831df46df"} Oct 04 03:30:10 crc kubenswrapper[4742]: I1004 03:30:10.874918 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e4971fb-098a-4711-a4f6-370b1a5a1120" path="/var/lib/kubelet/pods/3e4971fb-098a-4711-a4f6-370b1a5a1120/volumes" Oct 04 03:30:11 crc kubenswrapper[4742]: I1004 03:30:11.805483 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="7c3639a9-3e5f-44a6-9db3-1f60ca9ad489" containerName="glance-log" containerID="cri-o://fa19eaa8e6a0fba7017dde3d78e5ae7240f4ccac449bca91b170f44eb00f8ce1" gracePeriod=30 Oct 04 03:30:11 crc kubenswrapper[4742]: I1004 03:30:11.807649 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"7cad2cc9-60ac-4e19-81cc-4676387a2176","Type":"ContainerStarted","Data":"418b87a70546cf57fcad9e918164eb68b2723d7f74cd3a84eda8a614427ad686"} Oct 04 03:30:11 crc kubenswrapper[4742]: I1004 03:30:11.807687 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"7cad2cc9-60ac-4e19-81cc-4676387a2176","Type":"ContainerStarted","Data":"5b9ebe0007c311d7882fa5027411235e9643a698bd2d67537929fb64e601faeb"} Oct 04 03:30:11 crc kubenswrapper[4742]: I1004 03:30:11.808017 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="7c3639a9-3e5f-44a6-9db3-1f60ca9ad489" containerName="glance-api" containerID="cri-o://ff221474def26197448aede829b1e388fc8ddf74d90f7eb9d6bcaf18e436fe70" gracePeriod=30 Oct 04 03:30:11 crc kubenswrapper[4742]: I1004 03:30:11.808111 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="7c3639a9-3e5f-44a6-9db3-1f60ca9ad489" containerName="glance-httpd" containerID="cri-o://cfd40cf0b885b8df11a6da79255f67c0d823f46256e9556716c960b0e8e81ad4" gracePeriod=30 Oct 04 03:30:11 crc kubenswrapper[4742]: I1004 03:30:11.843779 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-0" podStartSLOduration=2.843748095 podStartE2EDuration="2.843748095s" podCreationTimestamp="2025-10-04 03:30:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:30:11.838030159 +0000 UTC m=+1177.770367425" watchObservedRunningTime="2025-10-04 03:30:11.843748095 +0000 UTC m=+1177.776085311" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.353356 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.448560 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-etc-nvme\") pod \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.448642 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbzcw\" (UniqueName: \"kubernetes.io/projected/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-kube-api-access-xbzcw\") pod \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.448682 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-httpd-run\") pod \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.448710 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.448741 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-lib-modules\") pod \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.448780 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-dev\") pod \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.448799 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.448824 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-var-locks-brick\") pod \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.448880 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-scripts\") pod \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.448952 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-etc-iscsi\") pod \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.448981 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-run\") pod \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.449056 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-sys\") pod \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.449078 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-config-data\") pod \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.449106 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-logs\") pod \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\" (UID: \"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489\") " Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.449842 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "7c3639a9-3e5f-44a6-9db3-1f60ca9ad489" (UID: "7c3639a9-3e5f-44a6-9db3-1f60ca9ad489"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.449863 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-logs" (OuterVolumeSpecName: "logs") pod "7c3639a9-3e5f-44a6-9db3-1f60ca9ad489" (UID: "7c3639a9-3e5f-44a6-9db3-1f60ca9ad489"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.451137 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "7c3639a9-3e5f-44a6-9db3-1f60ca9ad489" (UID: "7c3639a9-3e5f-44a6-9db3-1f60ca9ad489"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.451336 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-dev" (OuterVolumeSpecName: "dev") pod "7c3639a9-3e5f-44a6-9db3-1f60ca9ad489" (UID: "7c3639a9-3e5f-44a6-9db3-1f60ca9ad489"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.451334 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "7c3639a9-3e5f-44a6-9db3-1f60ca9ad489" (UID: "7c3639a9-3e5f-44a6-9db3-1f60ca9ad489"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.451432 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7c3639a9-3e5f-44a6-9db3-1f60ca9ad489" (UID: "7c3639a9-3e5f-44a6-9db3-1f60ca9ad489"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.451483 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-run" (OuterVolumeSpecName: "run") pod "7c3639a9-3e5f-44a6-9db3-1f60ca9ad489" (UID: "7c3639a9-3e5f-44a6-9db3-1f60ca9ad489"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.452026 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-sys" (OuterVolumeSpecName: "sys") pod "7c3639a9-3e5f-44a6-9db3-1f60ca9ad489" (UID: "7c3639a9-3e5f-44a6-9db3-1f60ca9ad489"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.452461 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "7c3639a9-3e5f-44a6-9db3-1f60ca9ad489" (UID: "7c3639a9-3e5f-44a6-9db3-1f60ca9ad489"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.491510 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance-cache") pod "7c3639a9-3e5f-44a6-9db3-1f60ca9ad489" (UID: "7c3639a9-3e5f-44a6-9db3-1f60ca9ad489"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.499872 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-scripts" (OuterVolumeSpecName: "scripts") pod "7c3639a9-3e5f-44a6-9db3-1f60ca9ad489" (UID: "7c3639a9-3e5f-44a6-9db3-1f60ca9ad489"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.510562 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "7c3639a9-3e5f-44a6-9db3-1f60ca9ad489" (UID: "7c3639a9-3e5f-44a6-9db3-1f60ca9ad489"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.511510 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-kube-api-access-xbzcw" (OuterVolumeSpecName: "kube-api-access-xbzcw") pod "7c3639a9-3e5f-44a6-9db3-1f60ca9ad489" (UID: "7c3639a9-3e5f-44a6-9db3-1f60ca9ad489"). InnerVolumeSpecName "kube-api-access-xbzcw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.552353 4742 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-sys\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.552401 4742 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-logs\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.552417 4742 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-etc-nvme\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.552431 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbzcw\" (UniqueName: \"kubernetes.io/projected/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-kube-api-access-xbzcw\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.552450 4742 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.552502 4742 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.552516 4742 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-lib-modules\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.552528 4742 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-dev\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.552545 4742 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.552560 4742 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-var-locks-brick\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.552571 4742 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.552582 4742 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-etc-iscsi\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.552593 4742 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-run\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.580945 4742 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.590117 4742 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.655153 4742 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.655188 4742 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.666209 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-config-data" (OuterVolumeSpecName: "config-data") pod "7c3639a9-3e5f-44a6-9db3-1f60ca9ad489" (UID: "7c3639a9-3e5f-44a6-9db3-1f60ca9ad489"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.756152 4742 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.818202 4742 generic.go:334] "Generic (PLEG): container finished" podID="7c3639a9-3e5f-44a6-9db3-1f60ca9ad489" containerID="ff221474def26197448aede829b1e388fc8ddf74d90f7eb9d6bcaf18e436fe70" exitCode=143 Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.818233 4742 generic.go:334] "Generic (PLEG): container finished" podID="7c3639a9-3e5f-44a6-9db3-1f60ca9ad489" containerID="cfd40cf0b885b8df11a6da79255f67c0d823f46256e9556716c960b0e8e81ad4" exitCode=0 Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.818244 4742 generic.go:334] "Generic (PLEG): container finished" podID="7c3639a9-3e5f-44a6-9db3-1f60ca9ad489" containerID="fa19eaa8e6a0fba7017dde3d78e5ae7240f4ccac449bca91b170f44eb00f8ce1" exitCode=143 Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.818287 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.818307 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489","Type":"ContainerDied","Data":"ff221474def26197448aede829b1e388fc8ddf74d90f7eb9d6bcaf18e436fe70"} Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.818374 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489","Type":"ContainerDied","Data":"cfd40cf0b885b8df11a6da79255f67c0d823f46256e9556716c960b0e8e81ad4"} Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.818387 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489","Type":"ContainerDied","Data":"fa19eaa8e6a0fba7017dde3d78e5ae7240f4ccac449bca91b170f44eb00f8ce1"} Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.818399 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"7c3639a9-3e5f-44a6-9db3-1f60ca9ad489","Type":"ContainerDied","Data":"c178df44d8da0b563563360cc55209fcee77fdecfcab82e34b5e8475bb36f7a8"} Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.818417 4742 scope.go:117] "RemoveContainer" containerID="ff221474def26197448aede829b1e388fc8ddf74d90f7eb9d6bcaf18e436fe70" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.846018 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.852836 4742 scope.go:117] "RemoveContainer" containerID="cfd40cf0b885b8df11a6da79255f67c0d823f46256e9556716c960b0e8e81ad4" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.886523 4742 scope.go:117] "RemoveContainer" containerID="fa19eaa8e6a0fba7017dde3d78e5ae7240f4ccac449bca91b170f44eb00f8ce1" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.900766 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.917366 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Oct 04 03:30:12 crc kubenswrapper[4742]: E1004 03:30:12.918286 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c3639a9-3e5f-44a6-9db3-1f60ca9ad489" containerName="glance-log" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.918304 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c3639a9-3e5f-44a6-9db3-1f60ca9ad489" containerName="glance-log" Oct 04 03:30:12 crc kubenswrapper[4742]: E1004 03:30:12.918386 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c3639a9-3e5f-44a6-9db3-1f60ca9ad489" containerName="glance-httpd" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.918394 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c3639a9-3e5f-44a6-9db3-1f60ca9ad489" containerName="glance-httpd" Oct 04 03:30:12 crc kubenswrapper[4742]: E1004 03:30:12.918410 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c3639a9-3e5f-44a6-9db3-1f60ca9ad489" containerName="glance-api" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.918417 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c3639a9-3e5f-44a6-9db3-1f60ca9ad489" containerName="glance-api" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.918783 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c3639a9-3e5f-44a6-9db3-1f60ca9ad489" containerName="glance-log" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.918819 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c3639a9-3e5f-44a6-9db3-1f60ca9ad489" containerName="glance-api" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.918855 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c3639a9-3e5f-44a6-9db3-1f60ca9ad489" containerName="glance-httpd" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.921705 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.931219 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.957196 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.983027 4742 scope.go:117] "RemoveContainer" containerID="ff221474def26197448aede829b1e388fc8ddf74d90f7eb9d6bcaf18e436fe70" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.984660 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/43cf2eff-93e3-4311-ac2f-ef3ecdfeb970-run\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.984718 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/43cf2eff-93e3-4311-ac2f-ef3ecdfeb970-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.984767 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.984783 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/43cf2eff-93e3-4311-ac2f-ef3ecdfeb970-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.984845 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43cf2eff-93e3-4311-ac2f-ef3ecdfeb970-scripts\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.984888 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/43cf2eff-93e3-4311-ac2f-ef3ecdfeb970-dev\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.984904 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/43cf2eff-93e3-4311-ac2f-ef3ecdfeb970-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.985000 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/43cf2eff-93e3-4311-ac2f-ef3ecdfeb970-sys\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.985051 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/43cf2eff-93e3-4311-ac2f-ef3ecdfeb970-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.985094 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.985132 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/43cf2eff-93e3-4311-ac2f-ef3ecdfeb970-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.985151 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwn75\" (UniqueName: \"kubernetes.io/projected/43cf2eff-93e3-4311-ac2f-ef3ecdfeb970-kube-api-access-gwn75\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.985179 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/43cf2eff-93e3-4311-ac2f-ef3ecdfeb970-logs\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.985204 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43cf2eff-93e3-4311-ac2f-ef3ecdfeb970-config-data\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:12 crc kubenswrapper[4742]: E1004 03:30:12.985983 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff221474def26197448aede829b1e388fc8ddf74d90f7eb9d6bcaf18e436fe70\": container with ID starting with ff221474def26197448aede829b1e388fc8ddf74d90f7eb9d6bcaf18e436fe70 not found: ID does not exist" containerID="ff221474def26197448aede829b1e388fc8ddf74d90f7eb9d6bcaf18e436fe70" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.986207 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff221474def26197448aede829b1e388fc8ddf74d90f7eb9d6bcaf18e436fe70"} err="failed to get container status \"ff221474def26197448aede829b1e388fc8ddf74d90f7eb9d6bcaf18e436fe70\": rpc error: code = NotFound desc = could not find container \"ff221474def26197448aede829b1e388fc8ddf74d90f7eb9d6bcaf18e436fe70\": container with ID starting with ff221474def26197448aede829b1e388fc8ddf74d90f7eb9d6bcaf18e436fe70 not found: ID does not exist" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.986243 4742 scope.go:117] "RemoveContainer" containerID="cfd40cf0b885b8df11a6da79255f67c0d823f46256e9556716c960b0e8e81ad4" Oct 04 03:30:12 crc kubenswrapper[4742]: E1004 03:30:12.986749 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfd40cf0b885b8df11a6da79255f67c0d823f46256e9556716c960b0e8e81ad4\": container with ID starting with cfd40cf0b885b8df11a6da79255f67c0d823f46256e9556716c960b0e8e81ad4 not found: ID does not exist" containerID="cfd40cf0b885b8df11a6da79255f67c0d823f46256e9556716c960b0e8e81ad4" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.986769 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfd40cf0b885b8df11a6da79255f67c0d823f46256e9556716c960b0e8e81ad4"} err="failed to get container status \"cfd40cf0b885b8df11a6da79255f67c0d823f46256e9556716c960b0e8e81ad4\": rpc error: code = NotFound desc = could not find container \"cfd40cf0b885b8df11a6da79255f67c0d823f46256e9556716c960b0e8e81ad4\": container with ID starting with cfd40cf0b885b8df11a6da79255f67c0d823f46256e9556716c960b0e8e81ad4 not found: ID does not exist" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.986782 4742 scope.go:117] "RemoveContainer" containerID="fa19eaa8e6a0fba7017dde3d78e5ae7240f4ccac449bca91b170f44eb00f8ce1" Oct 04 03:30:12 crc kubenswrapper[4742]: E1004 03:30:12.987063 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa19eaa8e6a0fba7017dde3d78e5ae7240f4ccac449bca91b170f44eb00f8ce1\": container with ID starting with fa19eaa8e6a0fba7017dde3d78e5ae7240f4ccac449bca91b170f44eb00f8ce1 not found: ID does not exist" containerID="fa19eaa8e6a0fba7017dde3d78e5ae7240f4ccac449bca91b170f44eb00f8ce1" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.987079 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa19eaa8e6a0fba7017dde3d78e5ae7240f4ccac449bca91b170f44eb00f8ce1"} err="failed to get container status \"fa19eaa8e6a0fba7017dde3d78e5ae7240f4ccac449bca91b170f44eb00f8ce1\": rpc error: code = NotFound desc = could not find container \"fa19eaa8e6a0fba7017dde3d78e5ae7240f4ccac449bca91b170f44eb00f8ce1\": container with ID starting with fa19eaa8e6a0fba7017dde3d78e5ae7240f4ccac449bca91b170f44eb00f8ce1 not found: ID does not exist" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.987090 4742 scope.go:117] "RemoveContainer" containerID="ff221474def26197448aede829b1e388fc8ddf74d90f7eb9d6bcaf18e436fe70" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.988295 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff221474def26197448aede829b1e388fc8ddf74d90f7eb9d6bcaf18e436fe70"} err="failed to get container status \"ff221474def26197448aede829b1e388fc8ddf74d90f7eb9d6bcaf18e436fe70\": rpc error: code = NotFound desc = could not find container \"ff221474def26197448aede829b1e388fc8ddf74d90f7eb9d6bcaf18e436fe70\": container with ID starting with ff221474def26197448aede829b1e388fc8ddf74d90f7eb9d6bcaf18e436fe70 not found: ID does not exist" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.988320 4742 scope.go:117] "RemoveContainer" containerID="cfd40cf0b885b8df11a6da79255f67c0d823f46256e9556716c960b0e8e81ad4" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.988794 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfd40cf0b885b8df11a6da79255f67c0d823f46256e9556716c960b0e8e81ad4"} err="failed to get container status \"cfd40cf0b885b8df11a6da79255f67c0d823f46256e9556716c960b0e8e81ad4\": rpc error: code = NotFound desc = could not find container \"cfd40cf0b885b8df11a6da79255f67c0d823f46256e9556716c960b0e8e81ad4\": container with ID starting with cfd40cf0b885b8df11a6da79255f67c0d823f46256e9556716c960b0e8e81ad4 not found: ID does not exist" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.988822 4742 scope.go:117] "RemoveContainer" containerID="fa19eaa8e6a0fba7017dde3d78e5ae7240f4ccac449bca91b170f44eb00f8ce1" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.989625 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa19eaa8e6a0fba7017dde3d78e5ae7240f4ccac449bca91b170f44eb00f8ce1"} err="failed to get container status \"fa19eaa8e6a0fba7017dde3d78e5ae7240f4ccac449bca91b170f44eb00f8ce1\": rpc error: code = NotFound desc = could not find container \"fa19eaa8e6a0fba7017dde3d78e5ae7240f4ccac449bca91b170f44eb00f8ce1\": container with ID starting with fa19eaa8e6a0fba7017dde3d78e5ae7240f4ccac449bca91b170f44eb00f8ce1 not found: ID does not exist" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.989665 4742 scope.go:117] "RemoveContainer" containerID="ff221474def26197448aede829b1e388fc8ddf74d90f7eb9d6bcaf18e436fe70" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.989949 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff221474def26197448aede829b1e388fc8ddf74d90f7eb9d6bcaf18e436fe70"} err="failed to get container status \"ff221474def26197448aede829b1e388fc8ddf74d90f7eb9d6bcaf18e436fe70\": rpc error: code = NotFound desc = could not find container \"ff221474def26197448aede829b1e388fc8ddf74d90f7eb9d6bcaf18e436fe70\": container with ID starting with ff221474def26197448aede829b1e388fc8ddf74d90f7eb9d6bcaf18e436fe70 not found: ID does not exist" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.989991 4742 scope.go:117] "RemoveContainer" containerID="cfd40cf0b885b8df11a6da79255f67c0d823f46256e9556716c960b0e8e81ad4" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.990227 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfd40cf0b885b8df11a6da79255f67c0d823f46256e9556716c960b0e8e81ad4"} err="failed to get container status \"cfd40cf0b885b8df11a6da79255f67c0d823f46256e9556716c960b0e8e81ad4\": rpc error: code = NotFound desc = could not find container \"cfd40cf0b885b8df11a6da79255f67c0d823f46256e9556716c960b0e8e81ad4\": container with ID starting with cfd40cf0b885b8df11a6da79255f67c0d823f46256e9556716c960b0e8e81ad4 not found: ID does not exist" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.990247 4742 scope.go:117] "RemoveContainer" containerID="fa19eaa8e6a0fba7017dde3d78e5ae7240f4ccac449bca91b170f44eb00f8ce1" Oct 04 03:30:12 crc kubenswrapper[4742]: I1004 03:30:12.990497 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa19eaa8e6a0fba7017dde3d78e5ae7240f4ccac449bca91b170f44eb00f8ce1"} err="failed to get container status \"fa19eaa8e6a0fba7017dde3d78e5ae7240f4ccac449bca91b170f44eb00f8ce1\": rpc error: code = NotFound desc = could not find container \"fa19eaa8e6a0fba7017dde3d78e5ae7240f4ccac449bca91b170f44eb00f8ce1\": container with ID starting with fa19eaa8e6a0fba7017dde3d78e5ae7240f4ccac449bca91b170f44eb00f8ce1 not found: ID does not exist" Oct 04 03:30:13 crc kubenswrapper[4742]: I1004 03:30:13.087365 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43cf2eff-93e3-4311-ac2f-ef3ecdfeb970-scripts\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:13 crc kubenswrapper[4742]: I1004 03:30:13.087417 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/43cf2eff-93e3-4311-ac2f-ef3ecdfeb970-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:13 crc kubenswrapper[4742]: I1004 03:30:13.087438 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/43cf2eff-93e3-4311-ac2f-ef3ecdfeb970-dev\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:13 crc kubenswrapper[4742]: I1004 03:30:13.087492 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/43cf2eff-93e3-4311-ac2f-ef3ecdfeb970-sys\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:13 crc kubenswrapper[4742]: I1004 03:30:13.087525 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/43cf2eff-93e3-4311-ac2f-ef3ecdfeb970-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:13 crc kubenswrapper[4742]: I1004 03:30:13.087553 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:13 crc kubenswrapper[4742]: I1004 03:30:13.087576 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/43cf2eff-93e3-4311-ac2f-ef3ecdfeb970-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:13 crc kubenswrapper[4742]: I1004 03:30:13.087596 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwn75\" (UniqueName: \"kubernetes.io/projected/43cf2eff-93e3-4311-ac2f-ef3ecdfeb970-kube-api-access-gwn75\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:13 crc kubenswrapper[4742]: I1004 03:30:13.087614 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/43cf2eff-93e3-4311-ac2f-ef3ecdfeb970-logs\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:13 crc kubenswrapper[4742]: I1004 03:30:13.087634 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43cf2eff-93e3-4311-ac2f-ef3ecdfeb970-config-data\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:13 crc kubenswrapper[4742]: I1004 03:30:13.087664 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/43cf2eff-93e3-4311-ac2f-ef3ecdfeb970-run\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:13 crc kubenswrapper[4742]: I1004 03:30:13.087687 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/43cf2eff-93e3-4311-ac2f-ef3ecdfeb970-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:13 crc kubenswrapper[4742]: I1004 03:30:13.087709 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:13 crc kubenswrapper[4742]: I1004 03:30:13.087728 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/43cf2eff-93e3-4311-ac2f-ef3ecdfeb970-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:13 crc kubenswrapper[4742]: I1004 03:30:13.087817 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/43cf2eff-93e3-4311-ac2f-ef3ecdfeb970-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:13 crc kubenswrapper[4742]: I1004 03:30:13.087922 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/43cf2eff-93e3-4311-ac2f-ef3ecdfeb970-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:13 crc kubenswrapper[4742]: I1004 03:30:13.088008 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/43cf2eff-93e3-4311-ac2f-ef3ecdfeb970-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:13 crc kubenswrapper[4742]: I1004 03:30:13.088057 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/43cf2eff-93e3-4311-ac2f-ef3ecdfeb970-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:13 crc kubenswrapper[4742]: I1004 03:30:13.088098 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/43cf2eff-93e3-4311-ac2f-ef3ecdfeb970-dev\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:13 crc kubenswrapper[4742]: I1004 03:30:13.088173 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/43cf2eff-93e3-4311-ac2f-ef3ecdfeb970-run\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:13 crc kubenswrapper[4742]: I1004 03:30:13.088235 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/43cf2eff-93e3-4311-ac2f-ef3ecdfeb970-sys\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:13 crc kubenswrapper[4742]: I1004 03:30:13.088392 4742 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:13 crc kubenswrapper[4742]: I1004 03:30:13.088398 4742 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") device mount path \"/mnt/openstack/pv12\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:13 crc kubenswrapper[4742]: I1004 03:30:13.088661 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/43cf2eff-93e3-4311-ac2f-ef3ecdfeb970-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:13 crc kubenswrapper[4742]: I1004 03:30:13.089018 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/43cf2eff-93e3-4311-ac2f-ef3ecdfeb970-logs\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:13 crc kubenswrapper[4742]: I1004 03:30:13.104505 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43cf2eff-93e3-4311-ac2f-ef3ecdfeb970-scripts\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:13 crc kubenswrapper[4742]: I1004 03:30:13.104822 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43cf2eff-93e3-4311-ac2f-ef3ecdfeb970-config-data\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:13 crc kubenswrapper[4742]: I1004 03:30:13.110847 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwn75\" (UniqueName: \"kubernetes.io/projected/43cf2eff-93e3-4311-ac2f-ef3ecdfeb970-kube-api-access-gwn75\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:13 crc kubenswrapper[4742]: I1004 03:30:13.111778 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:13 crc kubenswrapper[4742]: I1004 03:30:13.125640 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:13 crc kubenswrapper[4742]: I1004 03:30:13.290525 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:13 crc kubenswrapper[4742]: I1004 03:30:13.576097 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Oct 04 03:30:13 crc kubenswrapper[4742]: W1004 03:30:13.604191 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43cf2eff_93e3_4311_ac2f_ef3ecdfeb970.slice/crio-a6e65ea13ff01d4c7a79f426ff0f2aadaf825555f5bb6db19c9079b142f5c1ab WatchSource:0}: Error finding container a6e65ea13ff01d4c7a79f426ff0f2aadaf825555f5bb6db19c9079b142f5c1ab: Status 404 returned error can't find the container with id a6e65ea13ff01d4c7a79f426ff0f2aadaf825555f5bb6db19c9079b142f5c1ab Oct 04 03:30:13 crc kubenswrapper[4742]: I1004 03:30:13.838760 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970","Type":"ContainerStarted","Data":"a417f4c1d5a1c64fc97a1b4bafb0b4c6e297b51180112b8a81bc4c0fecf5a5ae"} Oct 04 03:30:13 crc kubenswrapper[4742]: I1004 03:30:13.838828 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970","Type":"ContainerStarted","Data":"a6e65ea13ff01d4c7a79f426ff0f2aadaf825555f5bb6db19c9079b142f5c1ab"} Oct 04 03:30:14 crc kubenswrapper[4742]: I1004 03:30:14.857231 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970","Type":"ContainerStarted","Data":"e6f289470a8edaca3cadbf406fd1914f07835c4dbe781cf0e924d8b2f820166a"} Oct 04 03:30:14 crc kubenswrapper[4742]: I1004 03:30:14.887180 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c3639a9-3e5f-44a6-9db3-1f60ca9ad489" path="/var/lib/kubelet/pods/7c3639a9-3e5f-44a6-9db3-1f60ca9ad489/volumes" Oct 04 03:30:14 crc kubenswrapper[4742]: I1004 03:30:14.889927 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"43cf2eff-93e3-4311-ac2f-ef3ecdfeb970","Type":"ContainerStarted","Data":"b43d817d77c3f712f12ce308f577e0136b714da60bfb7d133c40854bf5d6cdc3"} Oct 04 03:30:14 crc kubenswrapper[4742]: I1004 03:30:14.943036 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=2.9429939750000003 podStartE2EDuration="2.942993975s" podCreationTimestamp="2025-10-04 03:30:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:30:14.935115153 +0000 UTC m=+1180.867452369" watchObservedRunningTime="2025-10-04 03:30:14.942993975 +0000 UTC m=+1180.875331161" Oct 04 03:30:20 crc kubenswrapper[4742]: I1004 03:30:20.228247 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:20 crc kubenswrapper[4742]: I1004 03:30:20.229126 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:20 crc kubenswrapper[4742]: I1004 03:30:20.229143 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:20 crc kubenswrapper[4742]: I1004 03:30:20.264039 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:20 crc kubenswrapper[4742]: I1004 03:30:20.273456 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:20 crc kubenswrapper[4742]: I1004 03:30:20.277251 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:20 crc kubenswrapper[4742]: I1004 03:30:20.933775 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:20 crc kubenswrapper[4742]: I1004 03:30:20.934080 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:20 crc kubenswrapper[4742]: I1004 03:30:20.934156 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:20 crc kubenswrapper[4742]: I1004 03:30:20.958862 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:20 crc kubenswrapper[4742]: I1004 03:30:20.960569 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:20 crc kubenswrapper[4742]: I1004 03:30:20.963102 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 03:30:23 crc kubenswrapper[4742]: I1004 03:30:23.291100 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:23 crc kubenswrapper[4742]: I1004 03:30:23.291896 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:23 crc kubenswrapper[4742]: I1004 03:30:23.291911 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:23 crc kubenswrapper[4742]: I1004 03:30:23.337751 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:23 crc kubenswrapper[4742]: I1004 03:30:23.339412 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:23 crc kubenswrapper[4742]: I1004 03:30:23.363308 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:23 crc kubenswrapper[4742]: I1004 03:30:23.962204 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:23 crc kubenswrapper[4742]: I1004 03:30:23.962316 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:23 crc kubenswrapper[4742]: I1004 03:30:23.962341 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:23 crc kubenswrapper[4742]: I1004 03:30:23.976771 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:23 crc kubenswrapper[4742]: I1004 03:30:23.978085 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:23 crc kubenswrapper[4742]: I1004 03:30:23.981340 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 03:30:44 crc kubenswrapper[4742]: I1004 03:30:44.872224 4742 patch_prober.go:28] interesting pod/machine-config-daemon-qdt7v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:30:44 crc kubenswrapper[4742]: I1004 03:30:44.873690 4742 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:31:14 crc kubenswrapper[4742]: I1004 03:31:14.872102 4742 patch_prober.go:28] interesting pod/machine-config-daemon-qdt7v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:31:14 crc kubenswrapper[4742]: I1004 03:31:14.873125 4742 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:31:44 crc kubenswrapper[4742]: I1004 03:31:44.872088 4742 patch_prober.go:28] interesting pod/machine-config-daemon-qdt7v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:31:44 crc kubenswrapper[4742]: I1004 03:31:44.873331 4742 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:31:44 crc kubenswrapper[4742]: I1004 03:31:44.882013 4742 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" Oct 04 03:31:44 crc kubenswrapper[4742]: I1004 03:31:44.883209 4742 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"97c026eaa827ea7984bf1c754436a808d45b23a630beb167b6e8d3d8d75609e3"} pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 03:31:44 crc kubenswrapper[4742]: I1004 03:31:44.883361 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerName="machine-config-daemon" containerID="cri-o://97c026eaa827ea7984bf1c754436a808d45b23a630beb167b6e8d3d8d75609e3" gracePeriod=600 Oct 04 03:31:45 crc kubenswrapper[4742]: I1004 03:31:45.846342 4742 generic.go:334] "Generic (PLEG): container finished" podID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerID="97c026eaa827ea7984bf1c754436a808d45b23a630beb167b6e8d3d8d75609e3" exitCode=0 Oct 04 03:31:45 crc kubenswrapper[4742]: I1004 03:31:45.846441 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" event={"ID":"989ea247-7ac9-4ade-8dcb-2632fef4f83e","Type":"ContainerDied","Data":"97c026eaa827ea7984bf1c754436a808d45b23a630beb167b6e8d3d8d75609e3"} Oct 04 03:31:45 crc kubenswrapper[4742]: I1004 03:31:45.847078 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" event={"ID":"989ea247-7ac9-4ade-8dcb-2632fef4f83e","Type":"ContainerStarted","Data":"fe0e4c82469c3b0d2c42efade18b46e463b27548bdab066a609f2bba422a56a6"} Oct 04 03:31:45 crc kubenswrapper[4742]: I1004 03:31:45.847115 4742 scope.go:117] "RemoveContainer" containerID="85a4c342170ba17dd23c083b2ba6f556cfcde17e4bf847c7bc6921140b4883ad" Oct 04 03:33:35 crc kubenswrapper[4742]: I1004 03:33:35.471554 4742 scope.go:117] "RemoveContainer" containerID="b37287f10fe9c704fc8c003dbf0705cce72ee37d28e6ab44f4abe12d8af869f8" Oct 04 03:33:35 crc kubenswrapper[4742]: I1004 03:33:35.507686 4742 scope.go:117] "RemoveContainer" containerID="03e6a2189290bf48fb3bced0fbf093a73e0ee75e232c27ce5b02bc4b67d08c7d" Oct 04 03:33:35 crc kubenswrapper[4742]: I1004 03:33:35.540152 4742 scope.go:117] "RemoveContainer" containerID="2c5d6c5186fcc36d6be9f9b3c45e55eb3c383e9f0d1bae92257f78ab051b135f" Oct 04 03:34:14 crc kubenswrapper[4742]: I1004 03:34:14.872308 4742 patch_prober.go:28] interesting pod/machine-config-daemon-qdt7v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:34:14 crc kubenswrapper[4742]: I1004 03:34:14.873167 4742 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:34:35 crc kubenswrapper[4742]: I1004 03:34:35.632386 4742 scope.go:117] "RemoveContainer" containerID="3ef20e27d6e9a853995b31d1d16cac238881afbf10f611ee3824536fc009132c" Oct 04 03:34:44 crc kubenswrapper[4742]: I1004 03:34:44.872200 4742 patch_prober.go:28] interesting pod/machine-config-daemon-qdt7v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:34:44 crc kubenswrapper[4742]: I1004 03:34:44.873238 4742 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:34:58 crc kubenswrapper[4742]: I1004 03:34:58.220638 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9xzvl"] Oct 04 03:34:58 crc kubenswrapper[4742]: I1004 03:34:58.224641 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9xzvl" Oct 04 03:34:58 crc kubenswrapper[4742]: I1004 03:34:58.261230 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9xzvl"] Oct 04 03:34:58 crc kubenswrapper[4742]: I1004 03:34:58.320287 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnj47\" (UniqueName: \"kubernetes.io/projected/9fe001c5-3a2e-4458-b519-b314f3ccb891-kube-api-access-fnj47\") pod \"redhat-operators-9xzvl\" (UID: \"9fe001c5-3a2e-4458-b519-b314f3ccb891\") " pod="openshift-marketplace/redhat-operators-9xzvl" Oct 04 03:34:58 crc kubenswrapper[4742]: I1004 03:34:58.320412 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fe001c5-3a2e-4458-b519-b314f3ccb891-catalog-content\") pod \"redhat-operators-9xzvl\" (UID: \"9fe001c5-3a2e-4458-b519-b314f3ccb891\") " pod="openshift-marketplace/redhat-operators-9xzvl" Oct 04 03:34:58 crc kubenswrapper[4742]: I1004 03:34:58.320462 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fe001c5-3a2e-4458-b519-b314f3ccb891-utilities\") pod \"redhat-operators-9xzvl\" (UID: \"9fe001c5-3a2e-4458-b519-b314f3ccb891\") " pod="openshift-marketplace/redhat-operators-9xzvl" Oct 04 03:34:58 crc kubenswrapper[4742]: I1004 03:34:58.421712 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fe001c5-3a2e-4458-b519-b314f3ccb891-catalog-content\") pod \"redhat-operators-9xzvl\" (UID: \"9fe001c5-3a2e-4458-b519-b314f3ccb891\") " pod="openshift-marketplace/redhat-operators-9xzvl" Oct 04 03:34:58 crc kubenswrapper[4742]: I1004 03:34:58.421786 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fe001c5-3a2e-4458-b519-b314f3ccb891-utilities\") pod \"redhat-operators-9xzvl\" (UID: \"9fe001c5-3a2e-4458-b519-b314f3ccb891\") " pod="openshift-marketplace/redhat-operators-9xzvl" Oct 04 03:34:58 crc kubenswrapper[4742]: I1004 03:34:58.421855 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnj47\" (UniqueName: \"kubernetes.io/projected/9fe001c5-3a2e-4458-b519-b314f3ccb891-kube-api-access-fnj47\") pod \"redhat-operators-9xzvl\" (UID: \"9fe001c5-3a2e-4458-b519-b314f3ccb891\") " pod="openshift-marketplace/redhat-operators-9xzvl" Oct 04 03:34:58 crc kubenswrapper[4742]: I1004 03:34:58.422524 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fe001c5-3a2e-4458-b519-b314f3ccb891-catalog-content\") pod \"redhat-operators-9xzvl\" (UID: \"9fe001c5-3a2e-4458-b519-b314f3ccb891\") " pod="openshift-marketplace/redhat-operators-9xzvl" Oct 04 03:34:58 crc kubenswrapper[4742]: I1004 03:34:58.422586 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fe001c5-3a2e-4458-b519-b314f3ccb891-utilities\") pod \"redhat-operators-9xzvl\" (UID: \"9fe001c5-3a2e-4458-b519-b314f3ccb891\") " pod="openshift-marketplace/redhat-operators-9xzvl" Oct 04 03:34:58 crc kubenswrapper[4742]: I1004 03:34:58.460630 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnj47\" (UniqueName: \"kubernetes.io/projected/9fe001c5-3a2e-4458-b519-b314f3ccb891-kube-api-access-fnj47\") pod \"redhat-operators-9xzvl\" (UID: \"9fe001c5-3a2e-4458-b519-b314f3ccb891\") " pod="openshift-marketplace/redhat-operators-9xzvl" Oct 04 03:34:58 crc kubenswrapper[4742]: I1004 03:34:58.552081 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9xzvl" Oct 04 03:34:58 crc kubenswrapper[4742]: I1004 03:34:58.842329 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9xzvl"] Oct 04 03:34:59 crc kubenswrapper[4742]: I1004 03:34:59.068999 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9xzvl" event={"ID":"9fe001c5-3a2e-4458-b519-b314f3ccb891","Type":"ContainerStarted","Data":"3dab401040433eb55db78e45a30572d476ac2020d50ba3b0bfeb5e512cae28fe"} Oct 04 03:35:00 crc kubenswrapper[4742]: I1004 03:35:00.088680 4742 generic.go:334] "Generic (PLEG): container finished" podID="9fe001c5-3a2e-4458-b519-b314f3ccb891" containerID="007b58856720cf4e6dbe5c5e37cf9d02d21928dce17e408754d1a98c4b068231" exitCode=0 Oct 04 03:35:00 crc kubenswrapper[4742]: I1004 03:35:00.088954 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9xzvl" event={"ID":"9fe001c5-3a2e-4458-b519-b314f3ccb891","Type":"ContainerDied","Data":"007b58856720cf4e6dbe5c5e37cf9d02d21928dce17e408754d1a98c4b068231"} Oct 04 03:35:00 crc kubenswrapper[4742]: I1004 03:35:00.096316 4742 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 03:35:01 crc kubenswrapper[4742]: I1004 03:35:01.601524 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-g5xc9"] Oct 04 03:35:01 crc kubenswrapper[4742]: I1004 03:35:01.606580 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g5xc9" Oct 04 03:35:01 crc kubenswrapper[4742]: I1004 03:35:01.631748 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g5xc9"] Oct 04 03:35:01 crc kubenswrapper[4742]: I1004 03:35:01.788465 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7cef584a-1d0c-4810-8191-bc798cf0d4db-utilities\") pod \"certified-operators-g5xc9\" (UID: \"7cef584a-1d0c-4810-8191-bc798cf0d4db\") " pod="openshift-marketplace/certified-operators-g5xc9" Oct 04 03:35:01 crc kubenswrapper[4742]: I1004 03:35:01.788562 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7cef584a-1d0c-4810-8191-bc798cf0d4db-catalog-content\") pod \"certified-operators-g5xc9\" (UID: \"7cef584a-1d0c-4810-8191-bc798cf0d4db\") " pod="openshift-marketplace/certified-operators-g5xc9" Oct 04 03:35:01 crc kubenswrapper[4742]: I1004 03:35:01.788609 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tnc5\" (UniqueName: \"kubernetes.io/projected/7cef584a-1d0c-4810-8191-bc798cf0d4db-kube-api-access-7tnc5\") pod \"certified-operators-g5xc9\" (UID: \"7cef584a-1d0c-4810-8191-bc798cf0d4db\") " pod="openshift-marketplace/certified-operators-g5xc9" Oct 04 03:35:01 crc kubenswrapper[4742]: I1004 03:35:01.890593 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7cef584a-1d0c-4810-8191-bc798cf0d4db-utilities\") pod \"certified-operators-g5xc9\" (UID: \"7cef584a-1d0c-4810-8191-bc798cf0d4db\") " pod="openshift-marketplace/certified-operators-g5xc9" Oct 04 03:35:01 crc kubenswrapper[4742]: I1004 03:35:01.890665 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7cef584a-1d0c-4810-8191-bc798cf0d4db-catalog-content\") pod \"certified-operators-g5xc9\" (UID: \"7cef584a-1d0c-4810-8191-bc798cf0d4db\") " pod="openshift-marketplace/certified-operators-g5xc9" Oct 04 03:35:01 crc kubenswrapper[4742]: I1004 03:35:01.890701 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tnc5\" (UniqueName: \"kubernetes.io/projected/7cef584a-1d0c-4810-8191-bc798cf0d4db-kube-api-access-7tnc5\") pod \"certified-operators-g5xc9\" (UID: \"7cef584a-1d0c-4810-8191-bc798cf0d4db\") " pod="openshift-marketplace/certified-operators-g5xc9" Oct 04 03:35:01 crc kubenswrapper[4742]: I1004 03:35:01.891577 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7cef584a-1d0c-4810-8191-bc798cf0d4db-utilities\") pod \"certified-operators-g5xc9\" (UID: \"7cef584a-1d0c-4810-8191-bc798cf0d4db\") " pod="openshift-marketplace/certified-operators-g5xc9" Oct 04 03:35:01 crc kubenswrapper[4742]: I1004 03:35:01.891618 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7cef584a-1d0c-4810-8191-bc798cf0d4db-catalog-content\") pod \"certified-operators-g5xc9\" (UID: \"7cef584a-1d0c-4810-8191-bc798cf0d4db\") " pod="openshift-marketplace/certified-operators-g5xc9" Oct 04 03:35:01 crc kubenswrapper[4742]: I1004 03:35:01.926453 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tnc5\" (UniqueName: \"kubernetes.io/projected/7cef584a-1d0c-4810-8191-bc798cf0d4db-kube-api-access-7tnc5\") pod \"certified-operators-g5xc9\" (UID: \"7cef584a-1d0c-4810-8191-bc798cf0d4db\") " pod="openshift-marketplace/certified-operators-g5xc9" Oct 04 03:35:01 crc kubenswrapper[4742]: I1004 03:35:01.938137 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g5xc9" Oct 04 03:35:02 crc kubenswrapper[4742]: I1004 03:35:02.153351 4742 generic.go:334] "Generic (PLEG): container finished" podID="9fe001c5-3a2e-4458-b519-b314f3ccb891" containerID="929fb9286da3919b8353716cf7c759e06a2b0cace4579458987fb192acb844a2" exitCode=0 Oct 04 03:35:02 crc kubenswrapper[4742]: I1004 03:35:02.153405 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9xzvl" event={"ID":"9fe001c5-3a2e-4458-b519-b314f3ccb891","Type":"ContainerDied","Data":"929fb9286da3919b8353716cf7c759e06a2b0cace4579458987fb192acb844a2"} Oct 04 03:35:02 crc kubenswrapper[4742]: I1004 03:35:02.467320 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g5xc9"] Oct 04 03:35:02 crc kubenswrapper[4742]: W1004 03:35:02.477646 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7cef584a_1d0c_4810_8191_bc798cf0d4db.slice/crio-96d6b81ee68994e5cdd67db0fa4cf53a5a5830cda13dba1b34069fc8d535576c WatchSource:0}: Error finding container 96d6b81ee68994e5cdd67db0fa4cf53a5a5830cda13dba1b34069fc8d535576c: Status 404 returned error can't find the container with id 96d6b81ee68994e5cdd67db0fa4cf53a5a5830cda13dba1b34069fc8d535576c Oct 04 03:35:03 crc kubenswrapper[4742]: I1004 03:35:03.165406 4742 generic.go:334] "Generic (PLEG): container finished" podID="7cef584a-1d0c-4810-8191-bc798cf0d4db" containerID="09a3337c5c36901e05f462452e880263a61504e1b824954cf89f3d9a8a596e10" exitCode=0 Oct 04 03:35:03 crc kubenswrapper[4742]: I1004 03:35:03.165466 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g5xc9" event={"ID":"7cef584a-1d0c-4810-8191-bc798cf0d4db","Type":"ContainerDied","Data":"09a3337c5c36901e05f462452e880263a61504e1b824954cf89f3d9a8a596e10"} Oct 04 03:35:03 crc kubenswrapper[4742]: I1004 03:35:03.165842 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g5xc9" event={"ID":"7cef584a-1d0c-4810-8191-bc798cf0d4db","Type":"ContainerStarted","Data":"96d6b81ee68994e5cdd67db0fa4cf53a5a5830cda13dba1b34069fc8d535576c"} Oct 04 03:35:03 crc kubenswrapper[4742]: I1004 03:35:03.170227 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9xzvl" event={"ID":"9fe001c5-3a2e-4458-b519-b314f3ccb891","Type":"ContainerStarted","Data":"9c60d2aa781bb3fd065f624a969c980a78e04e3634af7a5864a3f54ec7e167b0"} Oct 04 03:35:03 crc kubenswrapper[4742]: I1004 03:35:03.227942 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9xzvl" podStartSLOduration=2.636052973 podStartE2EDuration="5.227897231s" podCreationTimestamp="2025-10-04 03:34:58 +0000 UTC" firstStartedPulling="2025-10-04 03:35:00.096018308 +0000 UTC m=+1466.028355484" lastFinishedPulling="2025-10-04 03:35:02.687862526 +0000 UTC m=+1468.620199742" observedRunningTime="2025-10-04 03:35:03.212747343 +0000 UTC m=+1469.145084529" watchObservedRunningTime="2025-10-04 03:35:03.227897231 +0000 UTC m=+1469.160234447" Oct 04 03:35:04 crc kubenswrapper[4742]: I1004 03:35:04.180337 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g5xc9" event={"ID":"7cef584a-1d0c-4810-8191-bc798cf0d4db","Type":"ContainerStarted","Data":"66d0cd2c6ab84ac020a771b76f8ca6680c761e48ddd36eb00c94c08cf5e902e8"} Oct 04 03:35:05 crc kubenswrapper[4742]: I1004 03:35:05.191983 4742 generic.go:334] "Generic (PLEG): container finished" podID="7cef584a-1d0c-4810-8191-bc798cf0d4db" containerID="66d0cd2c6ab84ac020a771b76f8ca6680c761e48ddd36eb00c94c08cf5e902e8" exitCode=0 Oct 04 03:35:05 crc kubenswrapper[4742]: I1004 03:35:05.192081 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g5xc9" event={"ID":"7cef584a-1d0c-4810-8191-bc798cf0d4db","Type":"ContainerDied","Data":"66d0cd2c6ab84ac020a771b76f8ca6680c761e48ddd36eb00c94c08cf5e902e8"} Oct 04 03:35:06 crc kubenswrapper[4742]: I1004 03:35:06.203374 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g5xc9" event={"ID":"7cef584a-1d0c-4810-8191-bc798cf0d4db","Type":"ContainerStarted","Data":"a1456191f4dd0301be4fc2f5941f334269f6c62f644f356187d8673519040020"} Oct 04 03:35:06 crc kubenswrapper[4742]: I1004 03:35:06.229922 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-g5xc9" podStartSLOduration=2.774083316 podStartE2EDuration="5.229904511s" podCreationTimestamp="2025-10-04 03:35:01 +0000 UTC" firstStartedPulling="2025-10-04 03:35:03.16768829 +0000 UTC m=+1469.100025496" lastFinishedPulling="2025-10-04 03:35:05.623509505 +0000 UTC m=+1471.555846691" observedRunningTime="2025-10-04 03:35:06.22833262 +0000 UTC m=+1472.160669796" watchObservedRunningTime="2025-10-04 03:35:06.229904511 +0000 UTC m=+1472.162241687" Oct 04 03:35:08 crc kubenswrapper[4742]: I1004 03:35:08.552628 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9xzvl" Oct 04 03:35:08 crc kubenswrapper[4742]: I1004 03:35:08.554598 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9xzvl" Oct 04 03:35:08 crc kubenswrapper[4742]: I1004 03:35:08.641079 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9xzvl" Oct 04 03:35:09 crc kubenswrapper[4742]: I1004 03:35:09.320201 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9xzvl" Oct 04 03:35:09 crc kubenswrapper[4742]: I1004 03:35:09.776923 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9xzvl"] Oct 04 03:35:11 crc kubenswrapper[4742]: I1004 03:35:11.258857 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9xzvl" podUID="9fe001c5-3a2e-4458-b519-b314f3ccb891" containerName="registry-server" containerID="cri-o://9c60d2aa781bb3fd065f624a969c980a78e04e3634af7a5864a3f54ec7e167b0" gracePeriod=2 Oct 04 03:35:11 crc kubenswrapper[4742]: I1004 03:35:11.807042 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9xzvl" Oct 04 03:35:11 crc kubenswrapper[4742]: I1004 03:35:11.891722 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fe001c5-3a2e-4458-b519-b314f3ccb891-utilities\") pod \"9fe001c5-3a2e-4458-b519-b314f3ccb891\" (UID: \"9fe001c5-3a2e-4458-b519-b314f3ccb891\") " Oct 04 03:35:11 crc kubenswrapper[4742]: I1004 03:35:11.891938 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fe001c5-3a2e-4458-b519-b314f3ccb891-catalog-content\") pod \"9fe001c5-3a2e-4458-b519-b314f3ccb891\" (UID: \"9fe001c5-3a2e-4458-b519-b314f3ccb891\") " Oct 04 03:35:11 crc kubenswrapper[4742]: I1004 03:35:11.892043 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnj47\" (UniqueName: \"kubernetes.io/projected/9fe001c5-3a2e-4458-b519-b314f3ccb891-kube-api-access-fnj47\") pod \"9fe001c5-3a2e-4458-b519-b314f3ccb891\" (UID: \"9fe001c5-3a2e-4458-b519-b314f3ccb891\") " Oct 04 03:35:11 crc kubenswrapper[4742]: I1004 03:35:11.894005 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fe001c5-3a2e-4458-b519-b314f3ccb891-utilities" (OuterVolumeSpecName: "utilities") pod "9fe001c5-3a2e-4458-b519-b314f3ccb891" (UID: "9fe001c5-3a2e-4458-b519-b314f3ccb891"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:35:11 crc kubenswrapper[4742]: I1004 03:35:11.894419 4742 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fe001c5-3a2e-4458-b519-b314f3ccb891-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:35:11 crc kubenswrapper[4742]: I1004 03:35:11.901223 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fe001c5-3a2e-4458-b519-b314f3ccb891-kube-api-access-fnj47" (OuterVolumeSpecName: "kube-api-access-fnj47") pod "9fe001c5-3a2e-4458-b519-b314f3ccb891" (UID: "9fe001c5-3a2e-4458-b519-b314f3ccb891"). InnerVolumeSpecName "kube-api-access-fnj47". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:35:11 crc kubenswrapper[4742]: I1004 03:35:11.938715 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-g5xc9" Oct 04 03:35:11 crc kubenswrapper[4742]: I1004 03:35:11.938882 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-g5xc9" Oct 04 03:35:11 crc kubenswrapper[4742]: I1004 03:35:11.997111 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnj47\" (UniqueName: \"kubernetes.io/projected/9fe001c5-3a2e-4458-b519-b314f3ccb891-kube-api-access-fnj47\") on node \"crc\" DevicePath \"\"" Oct 04 03:35:12 crc kubenswrapper[4742]: I1004 03:35:12.027522 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-g5xc9" Oct 04 03:35:12 crc kubenswrapper[4742]: I1004 03:35:12.273517 4742 generic.go:334] "Generic (PLEG): container finished" podID="9fe001c5-3a2e-4458-b519-b314f3ccb891" containerID="9c60d2aa781bb3fd065f624a969c980a78e04e3634af7a5864a3f54ec7e167b0" exitCode=0 Oct 04 03:35:12 crc kubenswrapper[4742]: I1004 03:35:12.273605 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9xzvl" event={"ID":"9fe001c5-3a2e-4458-b519-b314f3ccb891","Type":"ContainerDied","Data":"9c60d2aa781bb3fd065f624a969c980a78e04e3634af7a5864a3f54ec7e167b0"} Oct 04 03:35:12 crc kubenswrapper[4742]: I1004 03:35:12.274253 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9xzvl" event={"ID":"9fe001c5-3a2e-4458-b519-b314f3ccb891","Type":"ContainerDied","Data":"3dab401040433eb55db78e45a30572d476ac2020d50ba3b0bfeb5e512cae28fe"} Oct 04 03:35:12 crc kubenswrapper[4742]: I1004 03:35:12.273657 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9xzvl" Oct 04 03:35:12 crc kubenswrapper[4742]: I1004 03:35:12.274314 4742 scope.go:117] "RemoveContainer" containerID="9c60d2aa781bb3fd065f624a969c980a78e04e3634af7a5864a3f54ec7e167b0" Oct 04 03:35:12 crc kubenswrapper[4742]: I1004 03:35:12.300019 4742 scope.go:117] "RemoveContainer" containerID="929fb9286da3919b8353716cf7c759e06a2b0cace4579458987fb192acb844a2" Oct 04 03:35:12 crc kubenswrapper[4742]: I1004 03:35:12.336654 4742 scope.go:117] "RemoveContainer" containerID="007b58856720cf4e6dbe5c5e37cf9d02d21928dce17e408754d1a98c4b068231" Oct 04 03:35:12 crc kubenswrapper[4742]: I1004 03:35:12.367169 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-g5xc9" Oct 04 03:35:12 crc kubenswrapper[4742]: I1004 03:35:12.386091 4742 scope.go:117] "RemoveContainer" containerID="9c60d2aa781bb3fd065f624a969c980a78e04e3634af7a5864a3f54ec7e167b0" Oct 04 03:35:12 crc kubenswrapper[4742]: E1004 03:35:12.387202 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c60d2aa781bb3fd065f624a969c980a78e04e3634af7a5864a3f54ec7e167b0\": container with ID starting with 9c60d2aa781bb3fd065f624a969c980a78e04e3634af7a5864a3f54ec7e167b0 not found: ID does not exist" containerID="9c60d2aa781bb3fd065f624a969c980a78e04e3634af7a5864a3f54ec7e167b0" Oct 04 03:35:12 crc kubenswrapper[4742]: I1004 03:35:12.387245 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c60d2aa781bb3fd065f624a969c980a78e04e3634af7a5864a3f54ec7e167b0"} err="failed to get container status \"9c60d2aa781bb3fd065f624a969c980a78e04e3634af7a5864a3f54ec7e167b0\": rpc error: code = NotFound desc = could not find container \"9c60d2aa781bb3fd065f624a969c980a78e04e3634af7a5864a3f54ec7e167b0\": container with ID starting with 9c60d2aa781bb3fd065f624a969c980a78e04e3634af7a5864a3f54ec7e167b0 not found: ID does not exist" Oct 04 03:35:12 crc kubenswrapper[4742]: I1004 03:35:12.387296 4742 scope.go:117] "RemoveContainer" containerID="929fb9286da3919b8353716cf7c759e06a2b0cace4579458987fb192acb844a2" Oct 04 03:35:12 crc kubenswrapper[4742]: E1004 03:35:12.387815 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"929fb9286da3919b8353716cf7c759e06a2b0cace4579458987fb192acb844a2\": container with ID starting with 929fb9286da3919b8353716cf7c759e06a2b0cace4579458987fb192acb844a2 not found: ID does not exist" containerID="929fb9286da3919b8353716cf7c759e06a2b0cace4579458987fb192acb844a2" Oct 04 03:35:12 crc kubenswrapper[4742]: I1004 03:35:12.387927 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"929fb9286da3919b8353716cf7c759e06a2b0cace4579458987fb192acb844a2"} err="failed to get container status \"929fb9286da3919b8353716cf7c759e06a2b0cace4579458987fb192acb844a2\": rpc error: code = NotFound desc = could not find container \"929fb9286da3919b8353716cf7c759e06a2b0cace4579458987fb192acb844a2\": container with ID starting with 929fb9286da3919b8353716cf7c759e06a2b0cace4579458987fb192acb844a2 not found: ID does not exist" Oct 04 03:35:12 crc kubenswrapper[4742]: I1004 03:35:12.387979 4742 scope.go:117] "RemoveContainer" containerID="007b58856720cf4e6dbe5c5e37cf9d02d21928dce17e408754d1a98c4b068231" Oct 04 03:35:12 crc kubenswrapper[4742]: E1004 03:35:12.388534 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"007b58856720cf4e6dbe5c5e37cf9d02d21928dce17e408754d1a98c4b068231\": container with ID starting with 007b58856720cf4e6dbe5c5e37cf9d02d21928dce17e408754d1a98c4b068231 not found: ID does not exist" containerID="007b58856720cf4e6dbe5c5e37cf9d02d21928dce17e408754d1a98c4b068231" Oct 04 03:35:12 crc kubenswrapper[4742]: I1004 03:35:12.388617 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"007b58856720cf4e6dbe5c5e37cf9d02d21928dce17e408754d1a98c4b068231"} err="failed to get container status \"007b58856720cf4e6dbe5c5e37cf9d02d21928dce17e408754d1a98c4b068231\": rpc error: code = NotFound desc = could not find container \"007b58856720cf4e6dbe5c5e37cf9d02d21928dce17e408754d1a98c4b068231\": container with ID starting with 007b58856720cf4e6dbe5c5e37cf9d02d21928dce17e408754d1a98c4b068231 not found: ID does not exist" Oct 04 03:35:12 crc kubenswrapper[4742]: I1004 03:35:12.993768 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fe001c5-3a2e-4458-b519-b314f3ccb891-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9fe001c5-3a2e-4458-b519-b314f3ccb891" (UID: "9fe001c5-3a2e-4458-b519-b314f3ccb891"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:35:13 crc kubenswrapper[4742]: I1004 03:35:13.016420 4742 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fe001c5-3a2e-4458-b519-b314f3ccb891-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:35:13 crc kubenswrapper[4742]: I1004 03:35:13.177716 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-g5xc9"] Oct 04 03:35:13 crc kubenswrapper[4742]: I1004 03:35:13.228848 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9xzvl"] Oct 04 03:35:13 crc kubenswrapper[4742]: I1004 03:35:13.238571 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9xzvl"] Oct 04 03:35:14 crc kubenswrapper[4742]: I1004 03:35:14.297899 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-g5xc9" podUID="7cef584a-1d0c-4810-8191-bc798cf0d4db" containerName="registry-server" containerID="cri-o://a1456191f4dd0301be4fc2f5941f334269f6c62f644f356187d8673519040020" gracePeriod=2 Oct 04 03:35:14 crc kubenswrapper[4742]: I1004 03:35:14.870906 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fe001c5-3a2e-4458-b519-b314f3ccb891" path="/var/lib/kubelet/pods/9fe001c5-3a2e-4458-b519-b314f3ccb891/volumes" Oct 04 03:35:14 crc kubenswrapper[4742]: I1004 03:35:14.871769 4742 patch_prober.go:28] interesting pod/machine-config-daemon-qdt7v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:35:14 crc kubenswrapper[4742]: I1004 03:35:14.871881 4742 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:35:14 crc kubenswrapper[4742]: I1004 03:35:14.872838 4742 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" Oct 04 03:35:14 crc kubenswrapper[4742]: I1004 03:35:14.873661 4742 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fe0e4c82469c3b0d2c42efade18b46e463b27548bdab066a609f2bba422a56a6"} pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 03:35:14 crc kubenswrapper[4742]: I1004 03:35:14.873764 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerName="machine-config-daemon" containerID="cri-o://fe0e4c82469c3b0d2c42efade18b46e463b27548bdab066a609f2bba422a56a6" gracePeriod=600 Oct 04 03:35:15 crc kubenswrapper[4742]: I1004 03:35:15.309346 4742 generic.go:334] "Generic (PLEG): container finished" podID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerID="fe0e4c82469c3b0d2c42efade18b46e463b27548bdab066a609f2bba422a56a6" exitCode=0 Oct 04 03:35:15 crc kubenswrapper[4742]: I1004 03:35:15.309414 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" event={"ID":"989ea247-7ac9-4ade-8dcb-2632fef4f83e","Type":"ContainerDied","Data":"fe0e4c82469c3b0d2c42efade18b46e463b27548bdab066a609f2bba422a56a6"} Oct 04 03:35:15 crc kubenswrapper[4742]: I1004 03:35:15.309845 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" event={"ID":"989ea247-7ac9-4ade-8dcb-2632fef4f83e","Type":"ContainerStarted","Data":"ad8cd41abb89c3ef4bb7f33e14da9242af2dc910209a036b049d8fc459c07508"} Oct 04 03:35:15 crc kubenswrapper[4742]: I1004 03:35:15.309871 4742 scope.go:117] "RemoveContainer" containerID="97c026eaa827ea7984bf1c754436a808d45b23a630beb167b6e8d3d8d75609e3" Oct 04 03:35:15 crc kubenswrapper[4742]: I1004 03:35:15.313513 4742 generic.go:334] "Generic (PLEG): container finished" podID="7cef584a-1d0c-4810-8191-bc798cf0d4db" containerID="a1456191f4dd0301be4fc2f5941f334269f6c62f644f356187d8673519040020" exitCode=0 Oct 04 03:35:15 crc kubenswrapper[4742]: I1004 03:35:15.313556 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g5xc9" event={"ID":"7cef584a-1d0c-4810-8191-bc798cf0d4db","Type":"ContainerDied","Data":"a1456191f4dd0301be4fc2f5941f334269f6c62f644f356187d8673519040020"} Oct 04 03:35:15 crc kubenswrapper[4742]: I1004 03:35:15.313581 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g5xc9" event={"ID":"7cef584a-1d0c-4810-8191-bc798cf0d4db","Type":"ContainerDied","Data":"96d6b81ee68994e5cdd67db0fa4cf53a5a5830cda13dba1b34069fc8d535576c"} Oct 04 03:35:15 crc kubenswrapper[4742]: I1004 03:35:15.313597 4742 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96d6b81ee68994e5cdd67db0fa4cf53a5a5830cda13dba1b34069fc8d535576c" Oct 04 03:35:15 crc kubenswrapper[4742]: I1004 03:35:15.320389 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g5xc9" Oct 04 03:35:15 crc kubenswrapper[4742]: I1004 03:35:15.362979 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7cef584a-1d0c-4810-8191-bc798cf0d4db-catalog-content\") pod \"7cef584a-1d0c-4810-8191-bc798cf0d4db\" (UID: \"7cef584a-1d0c-4810-8191-bc798cf0d4db\") " Oct 04 03:35:15 crc kubenswrapper[4742]: I1004 03:35:15.363178 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7tnc5\" (UniqueName: \"kubernetes.io/projected/7cef584a-1d0c-4810-8191-bc798cf0d4db-kube-api-access-7tnc5\") pod \"7cef584a-1d0c-4810-8191-bc798cf0d4db\" (UID: \"7cef584a-1d0c-4810-8191-bc798cf0d4db\") " Oct 04 03:35:15 crc kubenswrapper[4742]: I1004 03:35:15.363262 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7cef584a-1d0c-4810-8191-bc798cf0d4db-utilities\") pod \"7cef584a-1d0c-4810-8191-bc798cf0d4db\" (UID: \"7cef584a-1d0c-4810-8191-bc798cf0d4db\") " Oct 04 03:35:15 crc kubenswrapper[4742]: I1004 03:35:15.364926 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7cef584a-1d0c-4810-8191-bc798cf0d4db-utilities" (OuterVolumeSpecName: "utilities") pod "7cef584a-1d0c-4810-8191-bc798cf0d4db" (UID: "7cef584a-1d0c-4810-8191-bc798cf0d4db"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:35:15 crc kubenswrapper[4742]: I1004 03:35:15.368280 4742 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7cef584a-1d0c-4810-8191-bc798cf0d4db-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:35:15 crc kubenswrapper[4742]: I1004 03:35:15.376958 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cef584a-1d0c-4810-8191-bc798cf0d4db-kube-api-access-7tnc5" (OuterVolumeSpecName: "kube-api-access-7tnc5") pod "7cef584a-1d0c-4810-8191-bc798cf0d4db" (UID: "7cef584a-1d0c-4810-8191-bc798cf0d4db"). InnerVolumeSpecName "kube-api-access-7tnc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:35:15 crc kubenswrapper[4742]: I1004 03:35:15.445678 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7cef584a-1d0c-4810-8191-bc798cf0d4db-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7cef584a-1d0c-4810-8191-bc798cf0d4db" (UID: "7cef584a-1d0c-4810-8191-bc798cf0d4db"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:35:15 crc kubenswrapper[4742]: I1004 03:35:15.469986 4742 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7cef584a-1d0c-4810-8191-bc798cf0d4db-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:35:15 crc kubenswrapper[4742]: I1004 03:35:15.470154 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7tnc5\" (UniqueName: \"kubernetes.io/projected/7cef584a-1d0c-4810-8191-bc798cf0d4db-kube-api-access-7tnc5\") on node \"crc\" DevicePath \"\"" Oct 04 03:35:16 crc kubenswrapper[4742]: I1004 03:35:16.330332 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g5xc9" Oct 04 03:35:16 crc kubenswrapper[4742]: I1004 03:35:16.382884 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-g5xc9"] Oct 04 03:35:16 crc kubenswrapper[4742]: I1004 03:35:16.396698 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-g5xc9"] Oct 04 03:35:16 crc kubenswrapper[4742]: I1004 03:35:16.875944 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cef584a-1d0c-4810-8191-bc798cf0d4db" path="/var/lib/kubelet/pods/7cef584a-1d0c-4810-8191-bc798cf0d4db/volumes" Oct 04 03:35:24 crc kubenswrapper[4742]: I1004 03:35:24.081306 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-db-create-vt8sv"] Oct 04 03:35:24 crc kubenswrapper[4742]: I1004 03:35:24.089616 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-db-create-vt8sv"] Oct 04 03:35:24 crc kubenswrapper[4742]: I1004 03:35:24.874742 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f409680-b99e-4ffb-bc3c-7f69c8856347" path="/var/lib/kubelet/pods/9f409680-b99e-4ffb-bc3c-7f69c8856347/volumes" Oct 04 03:35:31 crc kubenswrapper[4742]: I1004 03:35:31.810057 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2h2wx"] Oct 04 03:35:31 crc kubenswrapper[4742]: E1004 03:35:31.812004 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cef584a-1d0c-4810-8191-bc798cf0d4db" containerName="extract-utilities" Oct 04 03:35:31 crc kubenswrapper[4742]: I1004 03:35:31.812037 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cef584a-1d0c-4810-8191-bc798cf0d4db" containerName="extract-utilities" Oct 04 03:35:31 crc kubenswrapper[4742]: E1004 03:35:31.812170 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cef584a-1d0c-4810-8191-bc798cf0d4db" containerName="extract-content" Oct 04 03:35:31 crc kubenswrapper[4742]: I1004 03:35:31.812188 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cef584a-1d0c-4810-8191-bc798cf0d4db" containerName="extract-content" Oct 04 03:35:31 crc kubenswrapper[4742]: E1004 03:35:31.812237 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fe001c5-3a2e-4458-b519-b314f3ccb891" containerName="extract-utilities" Oct 04 03:35:31 crc kubenswrapper[4742]: I1004 03:35:31.812254 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fe001c5-3a2e-4458-b519-b314f3ccb891" containerName="extract-utilities" Oct 04 03:35:31 crc kubenswrapper[4742]: E1004 03:35:31.812308 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fe001c5-3a2e-4458-b519-b314f3ccb891" containerName="registry-server" Oct 04 03:35:31 crc kubenswrapper[4742]: I1004 03:35:31.812331 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fe001c5-3a2e-4458-b519-b314f3ccb891" containerName="registry-server" Oct 04 03:35:31 crc kubenswrapper[4742]: E1004 03:35:31.812393 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cef584a-1d0c-4810-8191-bc798cf0d4db" containerName="registry-server" Oct 04 03:35:31 crc kubenswrapper[4742]: I1004 03:35:31.812408 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cef584a-1d0c-4810-8191-bc798cf0d4db" containerName="registry-server" Oct 04 03:35:31 crc kubenswrapper[4742]: E1004 03:35:31.812455 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fe001c5-3a2e-4458-b519-b314f3ccb891" containerName="extract-content" Oct 04 03:35:31 crc kubenswrapper[4742]: I1004 03:35:31.812469 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fe001c5-3a2e-4458-b519-b314f3ccb891" containerName="extract-content" Oct 04 03:35:31 crc kubenswrapper[4742]: I1004 03:35:31.813084 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cef584a-1d0c-4810-8191-bc798cf0d4db" containerName="registry-server" Oct 04 03:35:31 crc kubenswrapper[4742]: I1004 03:35:31.813126 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fe001c5-3a2e-4458-b519-b314f3ccb891" containerName="registry-server" Oct 04 03:35:31 crc kubenswrapper[4742]: I1004 03:35:31.817464 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2h2wx" Oct 04 03:35:31 crc kubenswrapper[4742]: I1004 03:35:31.878143 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2h2wx"] Oct 04 03:35:31 crc kubenswrapper[4742]: I1004 03:35:31.899210 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d30942ba-e5d9-4e3f-a7f2-818c33e1f46b-utilities\") pod \"community-operators-2h2wx\" (UID: \"d30942ba-e5d9-4e3f-a7f2-818c33e1f46b\") " pod="openshift-marketplace/community-operators-2h2wx" Oct 04 03:35:31 crc kubenswrapper[4742]: I1004 03:35:31.899321 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d30942ba-e5d9-4e3f-a7f2-818c33e1f46b-catalog-content\") pod \"community-operators-2h2wx\" (UID: \"d30942ba-e5d9-4e3f-a7f2-818c33e1f46b\") " pod="openshift-marketplace/community-operators-2h2wx" Oct 04 03:35:31 crc kubenswrapper[4742]: I1004 03:35:31.899362 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwkhd\" (UniqueName: \"kubernetes.io/projected/d30942ba-e5d9-4e3f-a7f2-818c33e1f46b-kube-api-access-zwkhd\") pod \"community-operators-2h2wx\" (UID: \"d30942ba-e5d9-4e3f-a7f2-818c33e1f46b\") " pod="openshift-marketplace/community-operators-2h2wx" Oct 04 03:35:32 crc kubenswrapper[4742]: I1004 03:35:32.001898 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d30942ba-e5d9-4e3f-a7f2-818c33e1f46b-utilities\") pod \"community-operators-2h2wx\" (UID: \"d30942ba-e5d9-4e3f-a7f2-818c33e1f46b\") " pod="openshift-marketplace/community-operators-2h2wx" Oct 04 03:35:32 crc kubenswrapper[4742]: I1004 03:35:32.001994 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d30942ba-e5d9-4e3f-a7f2-818c33e1f46b-catalog-content\") pod \"community-operators-2h2wx\" (UID: \"d30942ba-e5d9-4e3f-a7f2-818c33e1f46b\") " pod="openshift-marketplace/community-operators-2h2wx" Oct 04 03:35:32 crc kubenswrapper[4742]: I1004 03:35:32.002027 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwkhd\" (UniqueName: \"kubernetes.io/projected/d30942ba-e5d9-4e3f-a7f2-818c33e1f46b-kube-api-access-zwkhd\") pod \"community-operators-2h2wx\" (UID: \"d30942ba-e5d9-4e3f-a7f2-818c33e1f46b\") " pod="openshift-marketplace/community-operators-2h2wx" Oct 04 03:35:32 crc kubenswrapper[4742]: I1004 03:35:32.002604 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d30942ba-e5d9-4e3f-a7f2-818c33e1f46b-utilities\") pod \"community-operators-2h2wx\" (UID: \"d30942ba-e5d9-4e3f-a7f2-818c33e1f46b\") " pod="openshift-marketplace/community-operators-2h2wx" Oct 04 03:35:32 crc kubenswrapper[4742]: I1004 03:35:32.003031 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d30942ba-e5d9-4e3f-a7f2-818c33e1f46b-catalog-content\") pod \"community-operators-2h2wx\" (UID: \"d30942ba-e5d9-4e3f-a7f2-818c33e1f46b\") " pod="openshift-marketplace/community-operators-2h2wx" Oct 04 03:35:32 crc kubenswrapper[4742]: I1004 03:35:32.021870 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwkhd\" (UniqueName: \"kubernetes.io/projected/d30942ba-e5d9-4e3f-a7f2-818c33e1f46b-kube-api-access-zwkhd\") pod \"community-operators-2h2wx\" (UID: \"d30942ba-e5d9-4e3f-a7f2-818c33e1f46b\") " pod="openshift-marketplace/community-operators-2h2wx" Oct 04 03:35:32 crc kubenswrapper[4742]: I1004 03:35:32.177709 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2h2wx" Oct 04 03:35:32 crc kubenswrapper[4742]: I1004 03:35:32.459548 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2h2wx"] Oct 04 03:35:32 crc kubenswrapper[4742]: I1004 03:35:32.528962 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2h2wx" event={"ID":"d30942ba-e5d9-4e3f-a7f2-818c33e1f46b","Type":"ContainerStarted","Data":"c909a14acb2ea3f3e3d6ed4a68f9353ec359f89a3a56b0319a85ddee2e90e54c"} Oct 04 03:35:33 crc kubenswrapper[4742]: I1004 03:35:33.543118 4742 generic.go:334] "Generic (PLEG): container finished" podID="d30942ba-e5d9-4e3f-a7f2-818c33e1f46b" containerID="74271f031297e4294ff64669dae090cb4db5c191611c19e59653af0689cfd266" exitCode=0 Oct 04 03:35:33 crc kubenswrapper[4742]: I1004 03:35:33.543164 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2h2wx" event={"ID":"d30942ba-e5d9-4e3f-a7f2-818c33e1f46b","Type":"ContainerDied","Data":"74271f031297e4294ff64669dae090cb4db5c191611c19e59653af0689cfd266"} Oct 04 03:35:34 crc kubenswrapper[4742]: I1004 03:35:34.054344 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-f46e-account-create-88lml"] Oct 04 03:35:34 crc kubenswrapper[4742]: I1004 03:35:34.066691 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-f46e-account-create-88lml"] Oct 04 03:35:34 crc kubenswrapper[4742]: I1004 03:35:34.557971 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2h2wx" event={"ID":"d30942ba-e5d9-4e3f-a7f2-818c33e1f46b","Type":"ContainerStarted","Data":"7bbcf3655a6f68689244f8782bd960d514b6a06d0501c8acbbe4c068d08cb887"} Oct 04 03:35:34 crc kubenswrapper[4742]: I1004 03:35:34.883994 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0af4aaa9-9a74-427f-9940-c5ebcf1d3efb" path="/var/lib/kubelet/pods/0af4aaa9-9a74-427f-9940-c5ebcf1d3efb/volumes" Oct 04 03:35:35 crc kubenswrapper[4742]: I1004 03:35:35.571559 4742 generic.go:334] "Generic (PLEG): container finished" podID="d30942ba-e5d9-4e3f-a7f2-818c33e1f46b" containerID="7bbcf3655a6f68689244f8782bd960d514b6a06d0501c8acbbe4c068d08cb887" exitCode=0 Oct 04 03:35:35 crc kubenswrapper[4742]: I1004 03:35:35.571794 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2h2wx" event={"ID":"d30942ba-e5d9-4e3f-a7f2-818c33e1f46b","Type":"ContainerDied","Data":"7bbcf3655a6f68689244f8782bd960d514b6a06d0501c8acbbe4c068d08cb887"} Oct 04 03:35:35 crc kubenswrapper[4742]: I1004 03:35:35.754421 4742 scope.go:117] "RemoveContainer" containerID="fc45ed162295579347770159149bc120feb68223e97a9ae51a67c5fb8ec22635" Oct 04 03:35:35 crc kubenswrapper[4742]: I1004 03:35:35.782647 4742 scope.go:117] "RemoveContainer" containerID="234ad6f29f016ac8c18cf7f2652ac3632eb27cdad39667dd4043b3cfcc1c4e8a" Oct 04 03:35:35 crc kubenswrapper[4742]: I1004 03:35:35.827238 4742 scope.go:117] "RemoveContainer" containerID="a1df9f270da043479282e7b102d2d7dc737edc597d3d0308489d98f5b11df8cf" Oct 04 03:35:35 crc kubenswrapper[4742]: I1004 03:35:35.858569 4742 scope.go:117] "RemoveContainer" containerID="ba39dafe40c73aa4f142a272463e0fc3da2dd8339ff54fd83a70b4cffc511d95" Oct 04 03:35:35 crc kubenswrapper[4742]: I1004 03:35:35.882094 4742 scope.go:117] "RemoveContainer" containerID="72d7cb4a52e36ca8b35f99369b884ceaf86641e02e37c43d3d81f74ad77f5e46" Oct 04 03:35:35 crc kubenswrapper[4742]: I1004 03:35:35.947900 4742 scope.go:117] "RemoveContainer" containerID="cd62b165b955c1d17f16d0bbfe0e35424a58f8197e129e653300ccce64a8fa31" Oct 04 03:35:36 crc kubenswrapper[4742]: I1004 03:35:36.585066 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2h2wx" event={"ID":"d30942ba-e5d9-4e3f-a7f2-818c33e1f46b","Type":"ContainerStarted","Data":"906eb3821a0f8cd0fc91bb8982b4d0a8ab867090b4a2f8950a82b802e85e2a35"} Oct 04 03:35:36 crc kubenswrapper[4742]: I1004 03:35:36.625191 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2h2wx" podStartSLOduration=3.168711608 podStartE2EDuration="5.625164009s" podCreationTimestamp="2025-10-04 03:35:31 +0000 UTC" firstStartedPulling="2025-10-04 03:35:33.546610411 +0000 UTC m=+1499.478947627" lastFinishedPulling="2025-10-04 03:35:36.003062852 +0000 UTC m=+1501.935400028" observedRunningTime="2025-10-04 03:35:36.614621129 +0000 UTC m=+1502.546958305" watchObservedRunningTime="2025-10-04 03:35:36.625164009 +0000 UTC m=+1502.557501215" Oct 04 03:35:42 crc kubenswrapper[4742]: I1004 03:35:42.179237 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2h2wx" Oct 04 03:35:42 crc kubenswrapper[4742]: I1004 03:35:42.180253 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2h2wx" Oct 04 03:35:42 crc kubenswrapper[4742]: I1004 03:35:42.238512 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2h2wx" Oct 04 03:35:42 crc kubenswrapper[4742]: I1004 03:35:42.726293 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2h2wx" Oct 04 03:35:42 crc kubenswrapper[4742]: I1004 03:35:42.801093 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2h2wx"] Oct 04 03:35:44 crc kubenswrapper[4742]: I1004 03:35:44.693635 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2h2wx" podUID="d30942ba-e5d9-4e3f-a7f2-818c33e1f46b" containerName="registry-server" containerID="cri-o://906eb3821a0f8cd0fc91bb8982b4d0a8ab867090b4a2f8950a82b802e85e2a35" gracePeriod=2 Oct 04 03:35:45 crc kubenswrapper[4742]: I1004 03:35:45.688802 4742 generic.go:334] "Generic (PLEG): container finished" podID="d30942ba-e5d9-4e3f-a7f2-818c33e1f46b" containerID="906eb3821a0f8cd0fc91bb8982b4d0a8ab867090b4a2f8950a82b802e85e2a35" exitCode=0 Oct 04 03:35:45 crc kubenswrapper[4742]: I1004 03:35:45.688883 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2h2wx" event={"ID":"d30942ba-e5d9-4e3f-a7f2-818c33e1f46b","Type":"ContainerDied","Data":"906eb3821a0f8cd0fc91bb8982b4d0a8ab867090b4a2f8950a82b802e85e2a35"} Oct 04 03:35:45 crc kubenswrapper[4742]: I1004 03:35:45.689228 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2h2wx" event={"ID":"d30942ba-e5d9-4e3f-a7f2-818c33e1f46b","Type":"ContainerDied","Data":"c909a14acb2ea3f3e3d6ed4a68f9353ec359f89a3a56b0319a85ddee2e90e54c"} Oct 04 03:35:45 crc kubenswrapper[4742]: I1004 03:35:45.689245 4742 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c909a14acb2ea3f3e3d6ed4a68f9353ec359f89a3a56b0319a85ddee2e90e54c" Oct 04 03:35:45 crc kubenswrapper[4742]: I1004 03:35:45.710290 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2h2wx" Oct 04 03:35:45 crc kubenswrapper[4742]: I1004 03:35:45.815723 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d30942ba-e5d9-4e3f-a7f2-818c33e1f46b-utilities\") pod \"d30942ba-e5d9-4e3f-a7f2-818c33e1f46b\" (UID: \"d30942ba-e5d9-4e3f-a7f2-818c33e1f46b\") " Oct 04 03:35:45 crc kubenswrapper[4742]: I1004 03:35:45.816134 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwkhd\" (UniqueName: \"kubernetes.io/projected/d30942ba-e5d9-4e3f-a7f2-818c33e1f46b-kube-api-access-zwkhd\") pod \"d30942ba-e5d9-4e3f-a7f2-818c33e1f46b\" (UID: \"d30942ba-e5d9-4e3f-a7f2-818c33e1f46b\") " Oct 04 03:35:45 crc kubenswrapper[4742]: I1004 03:35:45.816233 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d30942ba-e5d9-4e3f-a7f2-818c33e1f46b-catalog-content\") pod \"d30942ba-e5d9-4e3f-a7f2-818c33e1f46b\" (UID: \"d30942ba-e5d9-4e3f-a7f2-818c33e1f46b\") " Oct 04 03:35:45 crc kubenswrapper[4742]: I1004 03:35:45.818812 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d30942ba-e5d9-4e3f-a7f2-818c33e1f46b-utilities" (OuterVolumeSpecName: "utilities") pod "d30942ba-e5d9-4e3f-a7f2-818c33e1f46b" (UID: "d30942ba-e5d9-4e3f-a7f2-818c33e1f46b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:35:45 crc kubenswrapper[4742]: I1004 03:35:45.825503 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d30942ba-e5d9-4e3f-a7f2-818c33e1f46b-kube-api-access-zwkhd" (OuterVolumeSpecName: "kube-api-access-zwkhd") pod "d30942ba-e5d9-4e3f-a7f2-818c33e1f46b" (UID: "d30942ba-e5d9-4e3f-a7f2-818c33e1f46b"). InnerVolumeSpecName "kube-api-access-zwkhd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:35:45 crc kubenswrapper[4742]: I1004 03:35:45.870007 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d30942ba-e5d9-4e3f-a7f2-818c33e1f46b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d30942ba-e5d9-4e3f-a7f2-818c33e1f46b" (UID: "d30942ba-e5d9-4e3f-a7f2-818c33e1f46b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:35:45 crc kubenswrapper[4742]: I1004 03:35:45.917751 4742 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d30942ba-e5d9-4e3f-a7f2-818c33e1f46b-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:35:45 crc kubenswrapper[4742]: I1004 03:35:45.917791 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwkhd\" (UniqueName: \"kubernetes.io/projected/d30942ba-e5d9-4e3f-a7f2-818c33e1f46b-kube-api-access-zwkhd\") on node \"crc\" DevicePath \"\"" Oct 04 03:35:45 crc kubenswrapper[4742]: I1004 03:35:45.917810 4742 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d30942ba-e5d9-4e3f-a7f2-818c33e1f46b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:35:46 crc kubenswrapper[4742]: I1004 03:35:46.698767 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2h2wx" Oct 04 03:35:46 crc kubenswrapper[4742]: I1004 03:35:46.748544 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2h2wx"] Oct 04 03:35:46 crc kubenswrapper[4742]: I1004 03:35:46.757814 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2h2wx"] Oct 04 03:35:46 crc kubenswrapper[4742]: I1004 03:35:46.877012 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d30942ba-e5d9-4e3f-a7f2-818c33e1f46b" path="/var/lib/kubelet/pods/d30942ba-e5d9-4e3f-a7f2-818c33e1f46b/volumes" Oct 04 03:35:50 crc kubenswrapper[4742]: I1004 03:35:50.028844 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-j6wks"] Oct 04 03:35:50 crc kubenswrapper[4742]: I1004 03:35:50.035641 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-j6wks"] Oct 04 03:35:50 crc kubenswrapper[4742]: I1004 03:35:50.881250 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0778ac1c-afcb-4681-9622-29266b251908" path="/var/lib/kubelet/pods/0778ac1c-afcb-4681-9622-29266b251908/volumes" Oct 04 03:35:54 crc kubenswrapper[4742]: I1004 03:35:54.343566 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-95hn2"] Oct 04 03:35:54 crc kubenswrapper[4742]: E1004 03:35:54.344926 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d30942ba-e5d9-4e3f-a7f2-818c33e1f46b" containerName="extract-utilities" Oct 04 03:35:54 crc kubenswrapper[4742]: I1004 03:35:54.344948 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="d30942ba-e5d9-4e3f-a7f2-818c33e1f46b" containerName="extract-utilities" Oct 04 03:35:54 crc kubenswrapper[4742]: E1004 03:35:54.344978 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d30942ba-e5d9-4e3f-a7f2-818c33e1f46b" containerName="extract-content" Oct 04 03:35:54 crc kubenswrapper[4742]: I1004 03:35:54.344988 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="d30942ba-e5d9-4e3f-a7f2-818c33e1f46b" containerName="extract-content" Oct 04 03:35:54 crc kubenswrapper[4742]: E1004 03:35:54.345022 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d30942ba-e5d9-4e3f-a7f2-818c33e1f46b" containerName="registry-server" Oct 04 03:35:54 crc kubenswrapper[4742]: I1004 03:35:54.345034 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="d30942ba-e5d9-4e3f-a7f2-818c33e1f46b" containerName="registry-server" Oct 04 03:35:54 crc kubenswrapper[4742]: I1004 03:35:54.345311 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="d30942ba-e5d9-4e3f-a7f2-818c33e1f46b" containerName="registry-server" Oct 04 03:35:54 crc kubenswrapper[4742]: I1004 03:35:54.347042 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-95hn2" Oct 04 03:35:54 crc kubenswrapper[4742]: I1004 03:35:54.363611 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-95hn2"] Oct 04 03:35:54 crc kubenswrapper[4742]: I1004 03:35:54.491168 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8286157-8bec-4b3c-a22f-2c09e05912f3-utilities\") pod \"redhat-marketplace-95hn2\" (UID: \"d8286157-8bec-4b3c-a22f-2c09e05912f3\") " pod="openshift-marketplace/redhat-marketplace-95hn2" Oct 04 03:35:54 crc kubenswrapper[4742]: I1004 03:35:54.491238 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6z5x\" (UniqueName: \"kubernetes.io/projected/d8286157-8bec-4b3c-a22f-2c09e05912f3-kube-api-access-d6z5x\") pod \"redhat-marketplace-95hn2\" (UID: \"d8286157-8bec-4b3c-a22f-2c09e05912f3\") " pod="openshift-marketplace/redhat-marketplace-95hn2" Oct 04 03:35:54 crc kubenswrapper[4742]: I1004 03:35:54.491429 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8286157-8bec-4b3c-a22f-2c09e05912f3-catalog-content\") pod \"redhat-marketplace-95hn2\" (UID: \"d8286157-8bec-4b3c-a22f-2c09e05912f3\") " pod="openshift-marketplace/redhat-marketplace-95hn2" Oct 04 03:35:54 crc kubenswrapper[4742]: I1004 03:35:54.593645 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8286157-8bec-4b3c-a22f-2c09e05912f3-utilities\") pod \"redhat-marketplace-95hn2\" (UID: \"d8286157-8bec-4b3c-a22f-2c09e05912f3\") " pod="openshift-marketplace/redhat-marketplace-95hn2" Oct 04 03:35:54 crc kubenswrapper[4742]: I1004 03:35:54.594041 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6z5x\" (UniqueName: \"kubernetes.io/projected/d8286157-8bec-4b3c-a22f-2c09e05912f3-kube-api-access-d6z5x\") pod \"redhat-marketplace-95hn2\" (UID: \"d8286157-8bec-4b3c-a22f-2c09e05912f3\") " pod="openshift-marketplace/redhat-marketplace-95hn2" Oct 04 03:35:54 crc kubenswrapper[4742]: I1004 03:35:54.594355 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8286157-8bec-4b3c-a22f-2c09e05912f3-catalog-content\") pod \"redhat-marketplace-95hn2\" (UID: \"d8286157-8bec-4b3c-a22f-2c09e05912f3\") " pod="openshift-marketplace/redhat-marketplace-95hn2" Oct 04 03:35:54 crc kubenswrapper[4742]: I1004 03:35:54.594968 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8286157-8bec-4b3c-a22f-2c09e05912f3-utilities\") pod \"redhat-marketplace-95hn2\" (UID: \"d8286157-8bec-4b3c-a22f-2c09e05912f3\") " pod="openshift-marketplace/redhat-marketplace-95hn2" Oct 04 03:35:54 crc kubenswrapper[4742]: I1004 03:35:54.595184 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8286157-8bec-4b3c-a22f-2c09e05912f3-catalog-content\") pod \"redhat-marketplace-95hn2\" (UID: \"d8286157-8bec-4b3c-a22f-2c09e05912f3\") " pod="openshift-marketplace/redhat-marketplace-95hn2" Oct 04 03:35:54 crc kubenswrapper[4742]: I1004 03:35:54.624760 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6z5x\" (UniqueName: \"kubernetes.io/projected/d8286157-8bec-4b3c-a22f-2c09e05912f3-kube-api-access-d6z5x\") pod \"redhat-marketplace-95hn2\" (UID: \"d8286157-8bec-4b3c-a22f-2c09e05912f3\") " pod="openshift-marketplace/redhat-marketplace-95hn2" Oct 04 03:35:54 crc kubenswrapper[4742]: I1004 03:35:54.687077 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-95hn2" Oct 04 03:35:54 crc kubenswrapper[4742]: I1004 03:35:54.923430 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-95hn2"] Oct 04 03:35:55 crc kubenswrapper[4742]: I1004 03:35:55.834072 4742 generic.go:334] "Generic (PLEG): container finished" podID="d8286157-8bec-4b3c-a22f-2c09e05912f3" containerID="dbc98e9ceeda62a694271cf7d84e4a4c98f54d606bf0b8f56dc3fdc470564d88" exitCode=0 Oct 04 03:35:55 crc kubenswrapper[4742]: I1004 03:35:55.834470 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-95hn2" event={"ID":"d8286157-8bec-4b3c-a22f-2c09e05912f3","Type":"ContainerDied","Data":"dbc98e9ceeda62a694271cf7d84e4a4c98f54d606bf0b8f56dc3fdc470564d88"} Oct 04 03:35:55 crc kubenswrapper[4742]: I1004 03:35:55.834496 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-95hn2" event={"ID":"d8286157-8bec-4b3c-a22f-2c09e05912f3","Type":"ContainerStarted","Data":"e15e13cd804e4d35ee716a334723c6cb275106457da40639a82bd4413011b211"} Oct 04 03:35:56 crc kubenswrapper[4742]: I1004 03:35:56.845374 4742 generic.go:334] "Generic (PLEG): container finished" podID="d8286157-8bec-4b3c-a22f-2c09e05912f3" containerID="869f6cfc6b51d29a3c7fd503869c356d6fe9cccf27fd6defa67faec286075f8f" exitCode=0 Oct 04 03:35:56 crc kubenswrapper[4742]: I1004 03:35:56.845490 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-95hn2" event={"ID":"d8286157-8bec-4b3c-a22f-2c09e05912f3","Type":"ContainerDied","Data":"869f6cfc6b51d29a3c7fd503869c356d6fe9cccf27fd6defa67faec286075f8f"} Oct 04 03:35:57 crc kubenswrapper[4742]: I1004 03:35:57.031915 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-4rtpb"] Oct 04 03:35:57 crc kubenswrapper[4742]: I1004 03:35:57.039890 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-4rtpb"] Oct 04 03:35:57 crc kubenswrapper[4742]: I1004 03:35:57.878240 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-95hn2" event={"ID":"d8286157-8bec-4b3c-a22f-2c09e05912f3","Type":"ContainerStarted","Data":"326b815d4f56463451dfbaadbfaa4e328e16e212754738308144f6296a18f862"} Oct 04 03:35:57 crc kubenswrapper[4742]: I1004 03:35:57.904338 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-95hn2" podStartSLOduration=2.492962582 podStartE2EDuration="3.904313965s" podCreationTimestamp="2025-10-04 03:35:54 +0000 UTC" firstStartedPulling="2025-10-04 03:35:55.837941912 +0000 UTC m=+1521.770279088" lastFinishedPulling="2025-10-04 03:35:57.249293285 +0000 UTC m=+1523.181630471" observedRunningTime="2025-10-04 03:35:57.899635305 +0000 UTC m=+1523.831972491" watchObservedRunningTime="2025-10-04 03:35:57.904313965 +0000 UTC m=+1523.836651171" Oct 04 03:35:58 crc kubenswrapper[4742]: I1004 03:35:58.877064 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1593a5c-e6be-4abb-9d84-c78c90c8dcf6" path="/var/lib/kubelet/pods/d1593a5c-e6be-4abb-9d84-c78c90c8dcf6/volumes" Oct 04 03:36:04 crc kubenswrapper[4742]: I1004 03:36:04.687901 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-95hn2" Oct 04 03:36:04 crc kubenswrapper[4742]: I1004 03:36:04.689120 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-95hn2" Oct 04 03:36:04 crc kubenswrapper[4742]: I1004 03:36:04.772134 4742 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-95hn2" Oct 04 03:36:05 crc kubenswrapper[4742]: I1004 03:36:05.110262 4742 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-95hn2" Oct 04 03:36:05 crc kubenswrapper[4742]: I1004 03:36:05.191310 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-95hn2"] Oct 04 03:36:07 crc kubenswrapper[4742]: I1004 03:36:07.019619 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-95hn2" podUID="d8286157-8bec-4b3c-a22f-2c09e05912f3" containerName="registry-server" containerID="cri-o://326b815d4f56463451dfbaadbfaa4e328e16e212754738308144f6296a18f862" gracePeriod=2 Oct 04 03:36:08 crc kubenswrapper[4742]: I1004 03:36:08.034449 4742 generic.go:334] "Generic (PLEG): container finished" podID="d8286157-8bec-4b3c-a22f-2c09e05912f3" containerID="326b815d4f56463451dfbaadbfaa4e328e16e212754738308144f6296a18f862" exitCode=0 Oct 04 03:36:08 crc kubenswrapper[4742]: I1004 03:36:08.034546 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-95hn2" event={"ID":"d8286157-8bec-4b3c-a22f-2c09e05912f3","Type":"ContainerDied","Data":"326b815d4f56463451dfbaadbfaa4e328e16e212754738308144f6296a18f862"} Oct 04 03:36:08 crc kubenswrapper[4742]: I1004 03:36:08.128785 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-95hn2" Oct 04 03:36:08 crc kubenswrapper[4742]: I1004 03:36:08.288117 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8286157-8bec-4b3c-a22f-2c09e05912f3-catalog-content\") pod \"d8286157-8bec-4b3c-a22f-2c09e05912f3\" (UID: \"d8286157-8bec-4b3c-a22f-2c09e05912f3\") " Oct 04 03:36:08 crc kubenswrapper[4742]: I1004 03:36:08.288209 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6z5x\" (UniqueName: \"kubernetes.io/projected/d8286157-8bec-4b3c-a22f-2c09e05912f3-kube-api-access-d6z5x\") pod \"d8286157-8bec-4b3c-a22f-2c09e05912f3\" (UID: \"d8286157-8bec-4b3c-a22f-2c09e05912f3\") " Oct 04 03:36:08 crc kubenswrapper[4742]: I1004 03:36:08.288256 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8286157-8bec-4b3c-a22f-2c09e05912f3-utilities\") pod \"d8286157-8bec-4b3c-a22f-2c09e05912f3\" (UID: \"d8286157-8bec-4b3c-a22f-2c09e05912f3\") " Oct 04 03:36:08 crc kubenswrapper[4742]: I1004 03:36:08.289261 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8286157-8bec-4b3c-a22f-2c09e05912f3-utilities" (OuterVolumeSpecName: "utilities") pod "d8286157-8bec-4b3c-a22f-2c09e05912f3" (UID: "d8286157-8bec-4b3c-a22f-2c09e05912f3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:36:08 crc kubenswrapper[4742]: I1004 03:36:08.295547 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8286157-8bec-4b3c-a22f-2c09e05912f3-kube-api-access-d6z5x" (OuterVolumeSpecName: "kube-api-access-d6z5x") pod "d8286157-8bec-4b3c-a22f-2c09e05912f3" (UID: "d8286157-8bec-4b3c-a22f-2c09e05912f3"). InnerVolumeSpecName "kube-api-access-d6z5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:36:08 crc kubenswrapper[4742]: I1004 03:36:08.316458 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8286157-8bec-4b3c-a22f-2c09e05912f3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d8286157-8bec-4b3c-a22f-2c09e05912f3" (UID: "d8286157-8bec-4b3c-a22f-2c09e05912f3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:36:08 crc kubenswrapper[4742]: I1004 03:36:08.390714 4742 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8286157-8bec-4b3c-a22f-2c09e05912f3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:36:08 crc kubenswrapper[4742]: I1004 03:36:08.390783 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6z5x\" (UniqueName: \"kubernetes.io/projected/d8286157-8bec-4b3c-a22f-2c09e05912f3-kube-api-access-d6z5x\") on node \"crc\" DevicePath \"\"" Oct 04 03:36:08 crc kubenswrapper[4742]: I1004 03:36:08.390813 4742 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8286157-8bec-4b3c-a22f-2c09e05912f3-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:36:09 crc kubenswrapper[4742]: I1004 03:36:09.051388 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-95hn2" event={"ID":"d8286157-8bec-4b3c-a22f-2c09e05912f3","Type":"ContainerDied","Data":"e15e13cd804e4d35ee716a334723c6cb275106457da40639a82bd4413011b211"} Oct 04 03:36:09 crc kubenswrapper[4742]: I1004 03:36:09.051484 4742 scope.go:117] "RemoveContainer" containerID="326b815d4f56463451dfbaadbfaa4e328e16e212754738308144f6296a18f862" Oct 04 03:36:09 crc kubenswrapper[4742]: I1004 03:36:09.051512 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-95hn2" Oct 04 03:36:09 crc kubenswrapper[4742]: I1004 03:36:09.080184 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-95hn2"] Oct 04 03:36:09 crc kubenswrapper[4742]: I1004 03:36:09.092040 4742 scope.go:117] "RemoveContainer" containerID="869f6cfc6b51d29a3c7fd503869c356d6fe9cccf27fd6defa67faec286075f8f" Oct 04 03:36:09 crc kubenswrapper[4742]: I1004 03:36:09.095236 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-95hn2"] Oct 04 03:36:09 crc kubenswrapper[4742]: I1004 03:36:09.128286 4742 scope.go:117] "RemoveContainer" containerID="dbc98e9ceeda62a694271cf7d84e4a4c98f54d606bf0b8f56dc3fdc470564d88" Oct 04 03:36:10 crc kubenswrapper[4742]: I1004 03:36:10.877439 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8286157-8bec-4b3c-a22f-2c09e05912f3" path="/var/lib/kubelet/pods/d8286157-8bec-4b3c-a22f-2c09e05912f3/volumes" Oct 04 03:36:30 crc kubenswrapper[4742]: I1004 03:36:30.933060 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstackclient"] Oct 04 03:36:30 crc kubenswrapper[4742]: E1004 03:36:30.934503 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8286157-8bec-4b3c-a22f-2c09e05912f3" containerName="extract-content" Oct 04 03:36:30 crc kubenswrapper[4742]: I1004 03:36:30.934524 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8286157-8bec-4b3c-a22f-2c09e05912f3" containerName="extract-content" Oct 04 03:36:30 crc kubenswrapper[4742]: E1004 03:36:30.934550 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8286157-8bec-4b3c-a22f-2c09e05912f3" containerName="registry-server" Oct 04 03:36:30 crc kubenswrapper[4742]: I1004 03:36:30.934558 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8286157-8bec-4b3c-a22f-2c09e05912f3" containerName="registry-server" Oct 04 03:36:30 crc kubenswrapper[4742]: E1004 03:36:30.934572 4742 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8286157-8bec-4b3c-a22f-2c09e05912f3" containerName="extract-utilities" Oct 04 03:36:30 crc kubenswrapper[4742]: I1004 03:36:30.934581 4742 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8286157-8bec-4b3c-a22f-2c09e05912f3" containerName="extract-utilities" Oct 04 03:36:30 crc kubenswrapper[4742]: I1004 03:36:30.934762 4742 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8286157-8bec-4b3c-a22f-2c09e05912f3" containerName="registry-server" Oct 04 03:36:30 crc kubenswrapper[4742]: I1004 03:36:30.935488 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Oct 04 03:36:30 crc kubenswrapper[4742]: I1004 03:36:30.937544 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"openstack-config-secret" Oct 04 03:36:30 crc kubenswrapper[4742]: I1004 03:36:30.937745 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config" Oct 04 03:36:30 crc kubenswrapper[4742]: I1004 03:36:30.938092 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts-9db6gc427h" Oct 04 03:36:30 crc kubenswrapper[4742]: I1004 03:36:30.938011 4742 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"default-dockercfg-vcgns" Oct 04 03:36:30 crc kubenswrapper[4742]: I1004 03:36:30.943240 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Oct 04 03:36:31 crc kubenswrapper[4742]: I1004 03:36:31.042782 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/1bc82217-289b-4cff-9f9a-1b62bbfd03f6-openstack-scripts\") pod \"openstackclient\" (UID: \"1bc82217-289b-4cff-9f9a-1b62bbfd03f6\") " pod="glance-kuttl-tests/openstackclient" Oct 04 03:36:31 crc kubenswrapper[4742]: I1004 03:36:31.042850 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1bc82217-289b-4cff-9f9a-1b62bbfd03f6-openstack-config\") pod \"openstackclient\" (UID: \"1bc82217-289b-4cff-9f9a-1b62bbfd03f6\") " pod="glance-kuttl-tests/openstackclient" Oct 04 03:36:31 crc kubenswrapper[4742]: I1004 03:36:31.042878 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1bc82217-289b-4cff-9f9a-1b62bbfd03f6-openstack-config-secret\") pod \"openstackclient\" (UID: \"1bc82217-289b-4cff-9f9a-1b62bbfd03f6\") " pod="glance-kuttl-tests/openstackclient" Oct 04 03:36:31 crc kubenswrapper[4742]: I1004 03:36:31.042971 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kt7d\" (UniqueName: \"kubernetes.io/projected/1bc82217-289b-4cff-9f9a-1b62bbfd03f6-kube-api-access-2kt7d\") pod \"openstackclient\" (UID: \"1bc82217-289b-4cff-9f9a-1b62bbfd03f6\") " pod="glance-kuttl-tests/openstackclient" Oct 04 03:36:31 crc kubenswrapper[4742]: I1004 03:36:31.144575 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/1bc82217-289b-4cff-9f9a-1b62bbfd03f6-openstack-scripts\") pod \"openstackclient\" (UID: \"1bc82217-289b-4cff-9f9a-1b62bbfd03f6\") " pod="glance-kuttl-tests/openstackclient" Oct 04 03:36:31 crc kubenswrapper[4742]: I1004 03:36:31.144640 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1bc82217-289b-4cff-9f9a-1b62bbfd03f6-openstack-config\") pod \"openstackclient\" (UID: \"1bc82217-289b-4cff-9f9a-1b62bbfd03f6\") " pod="glance-kuttl-tests/openstackclient" Oct 04 03:36:31 crc kubenswrapper[4742]: I1004 03:36:31.144663 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1bc82217-289b-4cff-9f9a-1b62bbfd03f6-openstack-config-secret\") pod \"openstackclient\" (UID: \"1bc82217-289b-4cff-9f9a-1b62bbfd03f6\") " pod="glance-kuttl-tests/openstackclient" Oct 04 03:36:31 crc kubenswrapper[4742]: I1004 03:36:31.144753 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kt7d\" (UniqueName: \"kubernetes.io/projected/1bc82217-289b-4cff-9f9a-1b62bbfd03f6-kube-api-access-2kt7d\") pod \"openstackclient\" (UID: \"1bc82217-289b-4cff-9f9a-1b62bbfd03f6\") " pod="glance-kuttl-tests/openstackclient" Oct 04 03:36:31 crc kubenswrapper[4742]: I1004 03:36:31.145714 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/1bc82217-289b-4cff-9f9a-1b62bbfd03f6-openstack-scripts\") pod \"openstackclient\" (UID: \"1bc82217-289b-4cff-9f9a-1b62bbfd03f6\") " pod="glance-kuttl-tests/openstackclient" Oct 04 03:36:31 crc kubenswrapper[4742]: I1004 03:36:31.146513 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1bc82217-289b-4cff-9f9a-1b62bbfd03f6-openstack-config\") pod \"openstackclient\" (UID: \"1bc82217-289b-4cff-9f9a-1b62bbfd03f6\") " pod="glance-kuttl-tests/openstackclient" Oct 04 03:36:31 crc kubenswrapper[4742]: I1004 03:36:31.164440 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1bc82217-289b-4cff-9f9a-1b62bbfd03f6-openstack-config-secret\") pod \"openstackclient\" (UID: \"1bc82217-289b-4cff-9f9a-1b62bbfd03f6\") " pod="glance-kuttl-tests/openstackclient" Oct 04 03:36:31 crc kubenswrapper[4742]: I1004 03:36:31.165032 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kt7d\" (UniqueName: \"kubernetes.io/projected/1bc82217-289b-4cff-9f9a-1b62bbfd03f6-kube-api-access-2kt7d\") pod \"openstackclient\" (UID: \"1bc82217-289b-4cff-9f9a-1b62bbfd03f6\") " pod="glance-kuttl-tests/openstackclient" Oct 04 03:36:31 crc kubenswrapper[4742]: I1004 03:36:31.267572 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Oct 04 03:36:31 crc kubenswrapper[4742]: I1004 03:36:31.532311 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Oct 04 03:36:32 crc kubenswrapper[4742]: I1004 03:36:32.347017 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"1bc82217-289b-4cff-9f9a-1b62bbfd03f6","Type":"ContainerStarted","Data":"fe7463f070db881cd2621fc09417b256fe15370f77873d047f784abcab935904"} Oct 04 03:36:32 crc kubenswrapper[4742]: I1004 03:36:32.347779 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"1bc82217-289b-4cff-9f9a-1b62bbfd03f6","Type":"ContainerStarted","Data":"8743c24eab102db8aaf4692cb0aba127fb9cc2f494be9b7ffd00a8a5bbb83e8b"} Oct 04 03:36:32 crc kubenswrapper[4742]: I1004 03:36:32.387214 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstackclient" podStartSLOduration=2.387188859 podStartE2EDuration="2.387188859s" podCreationTimestamp="2025-10-04 03:36:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:36:32.380179396 +0000 UTC m=+1558.312516612" watchObservedRunningTime="2025-10-04 03:36:32.387188859 +0000 UTC m=+1558.319526065" Oct 04 03:36:36 crc kubenswrapper[4742]: I1004 03:36:36.149390 4742 scope.go:117] "RemoveContainer" containerID="f9228a893222be3d29903c9a506e9fa0d7b126e0f9ee0769bf64ad01241eeff8" Oct 04 03:36:36 crc kubenswrapper[4742]: I1004 03:36:36.220799 4742 scope.go:117] "RemoveContainer" containerID="b8cabab5b57e3720fdf1f8871b6ba110850453e8d3e7f0e7823b70a8470e6366" Oct 04 03:37:44 crc kubenswrapper[4742]: I1004 03:37:44.872322 4742 patch_prober.go:28] interesting pod/machine-config-daemon-qdt7v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:37:44 crc kubenswrapper[4742]: I1004 03:37:44.873322 4742 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:38:08 crc kubenswrapper[4742]: I1004 03:38:08.270901 4742 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-x49qd/must-gather-wcl4z"] Oct 04 03:38:08 crc kubenswrapper[4742]: I1004 03:38:08.273721 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x49qd/must-gather-wcl4z" Oct 04 03:38:08 crc kubenswrapper[4742]: I1004 03:38:08.274508 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b6afdf41-80ea-433f-924c-2627f10d96a4-must-gather-output\") pod \"must-gather-wcl4z\" (UID: \"b6afdf41-80ea-433f-924c-2627f10d96a4\") " pod="openshift-must-gather-x49qd/must-gather-wcl4z" Oct 04 03:38:08 crc kubenswrapper[4742]: I1004 03:38:08.274662 4742 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7h6kc\" (UniqueName: \"kubernetes.io/projected/b6afdf41-80ea-433f-924c-2627f10d96a4-kube-api-access-7h6kc\") pod \"must-gather-wcl4z\" (UID: \"b6afdf41-80ea-433f-924c-2627f10d96a4\") " pod="openshift-must-gather-x49qd/must-gather-wcl4z" Oct 04 03:38:08 crc kubenswrapper[4742]: I1004 03:38:08.276166 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-x49qd"/"openshift-service-ca.crt" Oct 04 03:38:08 crc kubenswrapper[4742]: I1004 03:38:08.276426 4742 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-x49qd"/"kube-root-ca.crt" Oct 04 03:38:08 crc kubenswrapper[4742]: I1004 03:38:08.292666 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-x49qd/must-gather-wcl4z"] Oct 04 03:38:08 crc kubenswrapper[4742]: I1004 03:38:08.391593 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b6afdf41-80ea-433f-924c-2627f10d96a4-must-gather-output\") pod \"must-gather-wcl4z\" (UID: \"b6afdf41-80ea-433f-924c-2627f10d96a4\") " pod="openshift-must-gather-x49qd/must-gather-wcl4z" Oct 04 03:38:08 crc kubenswrapper[4742]: I1004 03:38:08.391681 4742 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7h6kc\" (UniqueName: \"kubernetes.io/projected/b6afdf41-80ea-433f-924c-2627f10d96a4-kube-api-access-7h6kc\") pod \"must-gather-wcl4z\" (UID: \"b6afdf41-80ea-433f-924c-2627f10d96a4\") " pod="openshift-must-gather-x49qd/must-gather-wcl4z" Oct 04 03:38:08 crc kubenswrapper[4742]: I1004 03:38:08.392782 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b6afdf41-80ea-433f-924c-2627f10d96a4-must-gather-output\") pod \"must-gather-wcl4z\" (UID: \"b6afdf41-80ea-433f-924c-2627f10d96a4\") " pod="openshift-must-gather-x49qd/must-gather-wcl4z" Oct 04 03:38:08 crc kubenswrapper[4742]: I1004 03:38:08.414987 4742 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7h6kc\" (UniqueName: \"kubernetes.io/projected/b6afdf41-80ea-433f-924c-2627f10d96a4-kube-api-access-7h6kc\") pod \"must-gather-wcl4z\" (UID: \"b6afdf41-80ea-433f-924c-2627f10d96a4\") " pod="openshift-must-gather-x49qd/must-gather-wcl4z" Oct 04 03:38:08 crc kubenswrapper[4742]: I1004 03:38:08.655265 4742 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x49qd/must-gather-wcl4z" Oct 04 03:38:09 crc kubenswrapper[4742]: I1004 03:38:09.134044 4742 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-x49qd/must-gather-wcl4z"] Oct 04 03:38:09 crc kubenswrapper[4742]: W1004 03:38:09.150884 4742 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6afdf41_80ea_433f_924c_2627f10d96a4.slice/crio-2ae8e08014151d681af4c010911212ac36584a392ef725339ab07c1722c5f8f0 WatchSource:0}: Error finding container 2ae8e08014151d681af4c010911212ac36584a392ef725339ab07c1722c5f8f0: Status 404 returned error can't find the container with id 2ae8e08014151d681af4c010911212ac36584a392ef725339ab07c1722c5f8f0 Oct 04 03:38:09 crc kubenswrapper[4742]: I1004 03:38:09.407814 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x49qd/must-gather-wcl4z" event={"ID":"b6afdf41-80ea-433f-924c-2627f10d96a4","Type":"ContainerStarted","Data":"2ae8e08014151d681af4c010911212ac36584a392ef725339ab07c1722c5f8f0"} Oct 04 03:38:14 crc kubenswrapper[4742]: I1004 03:38:14.477036 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x49qd/must-gather-wcl4z" event={"ID":"b6afdf41-80ea-433f-924c-2627f10d96a4","Type":"ContainerStarted","Data":"242dca1004b99a0f29ad156d3edfe785a3c6ae74c39d5c96900b8bdd4866ba5e"} Oct 04 03:38:14 crc kubenswrapper[4742]: I1004 03:38:14.477896 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x49qd/must-gather-wcl4z" event={"ID":"b6afdf41-80ea-433f-924c-2627f10d96a4","Type":"ContainerStarted","Data":"936b3262f0bd1f4c1bd43d8b35714763a5281cf1c26244629b85efd9e87d39f7"} Oct 04 03:38:14 crc kubenswrapper[4742]: I1004 03:38:14.505148 4742 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-x49qd/must-gather-wcl4z" podStartSLOduration=2.416411748 podStartE2EDuration="6.505116646s" podCreationTimestamp="2025-10-04 03:38:08 +0000 UTC" firstStartedPulling="2025-10-04 03:38:09.158465175 +0000 UTC m=+1655.090802351" lastFinishedPulling="2025-10-04 03:38:13.247170073 +0000 UTC m=+1659.179507249" observedRunningTime="2025-10-04 03:38:14.502100427 +0000 UTC m=+1660.434437613" watchObservedRunningTime="2025-10-04 03:38:14.505116646 +0000 UTC m=+1660.437453852" Oct 04 03:38:14 crc kubenswrapper[4742]: I1004 03:38:14.871781 4742 patch_prober.go:28] interesting pod/machine-config-daemon-qdt7v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:38:14 crc kubenswrapper[4742]: I1004 03:38:14.871864 4742 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:38:44 crc kubenswrapper[4742]: I1004 03:38:44.871788 4742 patch_prober.go:28] interesting pod/machine-config-daemon-qdt7v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:38:44 crc kubenswrapper[4742]: I1004 03:38:44.872597 4742 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:38:44 crc kubenswrapper[4742]: I1004 03:38:44.872663 4742 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" Oct 04 03:38:44 crc kubenswrapper[4742]: I1004 03:38:44.873708 4742 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ad8cd41abb89c3ef4bb7f33e14da9242af2dc910209a036b049d8fc459c07508"} pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 03:38:44 crc kubenswrapper[4742]: I1004 03:38:44.873807 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerName="machine-config-daemon" containerID="cri-o://ad8cd41abb89c3ef4bb7f33e14da9242af2dc910209a036b049d8fc459c07508" gracePeriod=600 Oct 04 03:38:45 crc kubenswrapper[4742]: E1004 03:38:45.004764 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qdt7v_openshift-machine-config-operator(989ea247-7ac9-4ade-8dcb-2632fef4f83e)\"" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" Oct 04 03:38:45 crc kubenswrapper[4742]: I1004 03:38:45.800608 4742 generic.go:334] "Generic (PLEG): container finished" podID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" containerID="ad8cd41abb89c3ef4bb7f33e14da9242af2dc910209a036b049d8fc459c07508" exitCode=0 Oct 04 03:38:45 crc kubenswrapper[4742]: I1004 03:38:45.801027 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" event={"ID":"989ea247-7ac9-4ade-8dcb-2632fef4f83e","Type":"ContainerDied","Data":"ad8cd41abb89c3ef4bb7f33e14da9242af2dc910209a036b049d8fc459c07508"} Oct 04 03:38:45 crc kubenswrapper[4742]: I1004 03:38:45.801070 4742 scope.go:117] "RemoveContainer" containerID="fe0e4c82469c3b0d2c42efade18b46e463b27548bdab066a609f2bba422a56a6" Oct 04 03:38:45 crc kubenswrapper[4742]: I1004 03:38:45.802374 4742 scope.go:117] "RemoveContainer" containerID="ad8cd41abb89c3ef4bb7f33e14da9242af2dc910209a036b049d8fc459c07508" Oct 04 03:38:45 crc kubenswrapper[4742]: E1004 03:38:45.802926 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qdt7v_openshift-machine-config-operator(989ea247-7ac9-4ade-8dcb-2632fef4f83e)\"" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" Oct 04 03:38:53 crc kubenswrapper[4742]: I1004 03:38:53.204812 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7_c6aa45a1-997c-4038-97cb-12db2e9ca192/util/0.log" Oct 04 03:38:53 crc kubenswrapper[4742]: I1004 03:38:53.404536 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7_c6aa45a1-997c-4038-97cb-12db2e9ca192/pull/0.log" Oct 04 03:38:53 crc kubenswrapper[4742]: I1004 03:38:53.407120 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7_c6aa45a1-997c-4038-97cb-12db2e9ca192/pull/0.log" Oct 04 03:38:53 crc kubenswrapper[4742]: I1004 03:38:53.424952 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7_c6aa45a1-997c-4038-97cb-12db2e9ca192/util/0.log" Oct 04 03:38:53 crc kubenswrapper[4742]: I1004 03:38:53.632003 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7_c6aa45a1-997c-4038-97cb-12db2e9ca192/util/0.log" Oct 04 03:38:53 crc kubenswrapper[4742]: I1004 03:38:53.708322 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7_c6aa45a1-997c-4038-97cb-12db2e9ca192/pull/0.log" Oct 04 03:38:53 crc kubenswrapper[4742]: I1004 03:38:53.748666 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b9528132r2v7_c6aa45a1-997c-4038-97cb-12db2e9ca192/extract/0.log" Oct 04 03:38:53 crc kubenswrapper[4742]: I1004 03:38:53.828844 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf_7c778fde-b55e-42ae-89bb-90b6396651c4/util/0.log" Oct 04 03:38:54 crc kubenswrapper[4742]: I1004 03:38:54.012054 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf_7c778fde-b55e-42ae-89bb-90b6396651c4/util/0.log" Oct 04 03:38:54 crc kubenswrapper[4742]: I1004 03:38:54.036472 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf_7c778fde-b55e-42ae-89bb-90b6396651c4/pull/0.log" Oct 04 03:38:54 crc kubenswrapper[4742]: I1004 03:38:54.043951 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf_7c778fde-b55e-42ae-89bb-90b6396651c4/pull/0.log" Oct 04 03:38:54 crc kubenswrapper[4742]: I1004 03:38:54.233042 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf_7c778fde-b55e-42ae-89bb-90b6396651c4/pull/0.log" Oct 04 03:38:54 crc kubenswrapper[4742]: I1004 03:38:54.292193 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf_7c778fde-b55e-42ae-89bb-90b6396651c4/extract/0.log" Oct 04 03:38:54 crc kubenswrapper[4742]: I1004 03:38:54.297291 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590n9cpf_7c778fde-b55e-42ae-89bb-90b6396651c4/util/0.log" Oct 04 03:38:54 crc kubenswrapper[4742]: I1004 03:38:54.455596 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9_8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c/util/0.log" Oct 04 03:38:54 crc kubenswrapper[4742]: I1004 03:38:54.649075 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9_8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c/pull/0.log" Oct 04 03:38:54 crc kubenswrapper[4742]: I1004 03:38:54.656438 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9_8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c/pull/0.log" Oct 04 03:38:54 crc kubenswrapper[4742]: I1004 03:38:54.688798 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9_8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c/util/0.log" Oct 04 03:38:54 crc kubenswrapper[4742]: I1004 03:38:54.870580 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9_8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c/extract/0.log" Oct 04 03:38:54 crc kubenswrapper[4742]: I1004 03:38:54.884891 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9_8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c/util/0.log" Oct 04 03:38:54 crc kubenswrapper[4742]: I1004 03:38:54.911165 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c5e5af6bae87a00b305010eb22f0ca23f0da637e83beae679fcf4fcdf1lccs9_8cfaf6b2-4b9d-468e-b2e6-ed05fe298d7c/pull/0.log" Oct 04 03:38:55 crc kubenswrapper[4742]: I1004 03:38:55.076156 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl_fda100d8-ce31-46d2-bb94-2edb9aabe58e/util/0.log" Oct 04 03:38:55 crc kubenswrapper[4742]: I1004 03:38:55.228821 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl_fda100d8-ce31-46d2-bb94-2edb9aabe58e/util/0.log" Oct 04 03:38:55 crc kubenswrapper[4742]: I1004 03:38:55.247343 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl_fda100d8-ce31-46d2-bb94-2edb9aabe58e/pull/0.log" Oct 04 03:38:55 crc kubenswrapper[4742]: I1004 03:38:55.255970 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl_fda100d8-ce31-46d2-bb94-2edb9aabe58e/pull/0.log" Oct 04 03:38:55 crc kubenswrapper[4742]: I1004 03:38:55.457428 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl_fda100d8-ce31-46d2-bb94-2edb9aabe58e/util/0.log" Oct 04 03:38:55 crc kubenswrapper[4742]: I1004 03:38:55.464135 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl_fda100d8-ce31-46d2-bb94-2edb9aabe58e/extract/0.log" Oct 04 03:38:55 crc kubenswrapper[4742]: I1004 03:38:55.516866 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32xdhvl_fda100d8-ce31-46d2-bb94-2edb9aabe58e/pull/0.log" Oct 04 03:38:55 crc kubenswrapper[4742]: I1004 03:38:55.619429 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8_639d681e-3475-410c-b738-6f3be2be5b4b/util/0.log" Oct 04 03:38:55 crc kubenswrapper[4742]: I1004 03:38:55.843089 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8_639d681e-3475-410c-b738-6f3be2be5b4b/util/0.log" Oct 04 03:38:55 crc kubenswrapper[4742]: I1004 03:38:55.849755 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8_639d681e-3475-410c-b738-6f3be2be5b4b/pull/0.log" Oct 04 03:38:55 crc kubenswrapper[4742]: I1004 03:38:55.898965 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8_639d681e-3475-410c-b738-6f3be2be5b4b/pull/0.log" Oct 04 03:38:56 crc kubenswrapper[4742]: I1004 03:38:56.064236 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8_639d681e-3475-410c-b738-6f3be2be5b4b/pull/0.log" Oct 04 03:38:56 crc kubenswrapper[4742]: I1004 03:38:56.068579 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8_639d681e-3475-410c-b738-6f3be2be5b4b/util/0.log" Oct 04 03:38:56 crc kubenswrapper[4742]: I1004 03:38:56.073110 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d93b99dddc714b0f4b2148f40016b9ead21cc18743d58ffe812e1bd436j72v8_639d681e-3475-410c-b738-6f3be2be5b4b/extract/0.log" Oct 04 03:38:56 crc kubenswrapper[4742]: I1004 03:38:56.085957 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr_d72194dd-00e3-474f-a278-35ec718e1fd6/util/0.log" Oct 04 03:38:56 crc kubenswrapper[4742]: I1004 03:38:56.278489 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr_d72194dd-00e3-474f-a278-35ec718e1fd6/util/0.log" Oct 04 03:38:56 crc kubenswrapper[4742]: I1004 03:38:56.283358 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr_d72194dd-00e3-474f-a278-35ec718e1fd6/pull/0.log" Oct 04 03:38:56 crc kubenswrapper[4742]: I1004 03:38:56.290971 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr_d72194dd-00e3-474f-a278-35ec718e1fd6/pull/0.log" Oct 04 03:38:56 crc kubenswrapper[4742]: I1004 03:38:56.440314 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr_d72194dd-00e3-474f-a278-35ec718e1fd6/pull/0.log" Oct 04 03:38:56 crc kubenswrapper[4742]: I1004 03:38:56.458880 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr_d72194dd-00e3-474f-a278-35ec718e1fd6/util/0.log" Oct 04 03:38:56 crc kubenswrapper[4742]: I1004 03:38:56.486529 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce7304025kldr_d72194dd-00e3-474f-a278-35ec718e1fd6/extract/0.log" Oct 04 03:38:56 crc kubenswrapper[4742]: I1004 03:38:56.516086 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq_b62bb235-84cf-41e7-806e-a3f1f03ee01b/util/0.log" Oct 04 03:38:56 crc kubenswrapper[4742]: I1004 03:38:56.703502 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq_b62bb235-84cf-41e7-806e-a3f1f03ee01b/pull/0.log" Oct 04 03:38:56 crc kubenswrapper[4742]: I1004 03:38:56.710184 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq_b62bb235-84cf-41e7-806e-a3f1f03ee01b/util/0.log" Oct 04 03:38:56 crc kubenswrapper[4742]: I1004 03:38:56.728331 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq_b62bb235-84cf-41e7-806e-a3f1f03ee01b/pull/0.log" Oct 04 03:38:56 crc kubenswrapper[4742]: I1004 03:38:56.874126 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq_b62bb235-84cf-41e7-806e-a3f1f03ee01b/pull/0.log" Oct 04 03:38:56 crc kubenswrapper[4742]: I1004 03:38:56.915895 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq_b62bb235-84cf-41e7-806e-a3f1f03ee01b/util/0.log" Oct 04 03:38:56 crc kubenswrapper[4742]: I1004 03:38:56.936426 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f038faa1e879443b04fe3f8f631a30d7dfa99960e4d0468b17dcfc81edvn4nq_b62bb235-84cf-41e7-806e-a3f1f03ee01b/extract/0.log" Oct 04 03:38:56 crc kubenswrapper[4742]: I1004 03:38:56.950643 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-67c6d4bf5f-ngkxb_94b82631-9d81-4926-8677-1d9addae8808/kube-rbac-proxy/0.log" Oct 04 03:38:57 crc kubenswrapper[4742]: I1004 03:38:57.151045 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-67c6d4bf5f-ngkxb_94b82631-9d81-4926-8677-1d9addae8808/manager/0.log" Oct 04 03:38:57 crc kubenswrapper[4742]: I1004 03:38:57.165379 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-index-dfskt_9f5106a7-ade4-48c9-9b96-95db975f8a99/registry-server/0.log" Oct 04 03:38:57 crc kubenswrapper[4742]: I1004 03:38:57.211976 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5994b8459f-9phsb_ab3ea796-e1d2-4d21-980c-648daedc36cd/kube-rbac-proxy/0.log" Oct 04 03:38:57 crc kubenswrapper[4742]: I1004 03:38:57.306595 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5994b8459f-9phsb_ab3ea796-e1d2-4d21-980c-648daedc36cd/manager/0.log" Oct 04 03:38:57 crc kubenswrapper[4742]: I1004 03:38:57.347044 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-index-2sn2g_d574cbce-5f5f-4093-8e44-4bd06c8c9eae/registry-server/0.log" Oct 04 03:38:57 crc kubenswrapper[4742]: I1004 03:38:57.392688 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-796fff894f-4ns75_533d4297-c693-4bd3-ad07-1a2d59aaa1e1/kube-rbac-proxy/0.log" Oct 04 03:38:57 crc kubenswrapper[4742]: I1004 03:38:57.506544 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-796fff894f-4ns75_533d4297-c693-4bd3-ad07-1a2d59aaa1e1/manager/0.log" Oct 04 03:38:57 crc kubenswrapper[4742]: I1004 03:38:57.580956 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-index-rlwxl_1ba2be5e-e2e5-45db-a77e-cbad59fee6a3/registry-server/0.log" Oct 04 03:38:57 crc kubenswrapper[4742]: I1004 03:38:57.626470 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-9d6cf7d55-znns7_07343d4b-0199-4669-9e2b-9eb697e8b387/kube-rbac-proxy/0.log" Oct 04 03:38:57 crc kubenswrapper[4742]: I1004 03:38:57.718955 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-9d6cf7d55-znns7_07343d4b-0199-4669-9e2b-9eb697e8b387/manager/0.log" Oct 04 03:38:57 crc kubenswrapper[4742]: I1004 03:38:57.766241 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-index-kbkhk_ca9efb6f-8e2d-43bc-ac51-a8673024fe54/registry-server/0.log" Oct 04 03:38:57 crc kubenswrapper[4742]: I1004 03:38:57.913579 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-86bc99bbf9-j65wm_e298bed2-0909-45f9-b422-21e118fc389a/kube-rbac-proxy/0.log" Oct 04 03:38:58 crc kubenswrapper[4742]: I1004 03:38:58.015514 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-86bc99bbf9-j65wm_e298bed2-0909-45f9-b422-21e118fc389a/manager/0.log" Oct 04 03:38:58 crc kubenswrapper[4742]: I1004 03:38:58.096245 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-index-s9z66_5cc2c58e-80c3-48e8-a5f8-3fc7a8326232/registry-server/0.log" Oct 04 03:38:58 crc kubenswrapper[4742]: I1004 03:38:58.149060 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-779fc9694b-8c4x6_b846aad4-0d5d-4023-bed1-59da3710e48a/operator/0.log" Oct 04 03:38:58 crc kubenswrapper[4742]: I1004 03:38:58.253770 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-index-nbgzv_3b222ee4-424b-4962-9722-6ab60c494c6f/registry-server/0.log" Oct 04 03:38:58 crc kubenswrapper[4742]: I1004 03:38:58.322507 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f47cfc687-xl4hq_744a78c0-6d93-4a39-bb0c-68f80ff9716b/kube-rbac-proxy/0.log" Oct 04 03:38:58 crc kubenswrapper[4742]: I1004 03:38:58.349704 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f47cfc687-xl4hq_744a78c0-6d93-4a39-bb0c-68f80ff9716b/manager/0.log" Oct 04 03:38:58 crc kubenswrapper[4742]: I1004 03:38:58.453947 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-index-gppbc_a7fdeeb5-30fc-42da-b8c6-f5e91634cf91/registry-server/0.log" Oct 04 03:39:00 crc kubenswrapper[4742]: I1004 03:39:00.859046 4742 scope.go:117] "RemoveContainer" containerID="ad8cd41abb89c3ef4bb7f33e14da9242af2dc910209a036b049d8fc459c07508" Oct 04 03:39:00 crc kubenswrapper[4742]: E1004 03:39:00.860444 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qdt7v_openshift-machine-config-operator(989ea247-7ac9-4ade-8dcb-2632fef4f83e)\"" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" Oct 04 03:39:12 crc kubenswrapper[4742]: I1004 03:39:12.859422 4742 scope.go:117] "RemoveContainer" containerID="ad8cd41abb89c3ef4bb7f33e14da9242af2dc910209a036b049d8fc459c07508" Oct 04 03:39:12 crc kubenswrapper[4742]: E1004 03:39:12.860273 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qdt7v_openshift-machine-config-operator(989ea247-7ac9-4ade-8dcb-2632fef4f83e)\"" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" Oct 04 03:39:14 crc kubenswrapper[4742]: I1004 03:39:14.548912 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-ncmp2_62ad7c0a-a108-4c0f-b998-1accb11feb51/control-plane-machine-set-operator/0.log" Oct 04 03:39:14 crc kubenswrapper[4742]: I1004 03:39:14.750592 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-fcc5d_0062ea87-e9b5-44f9-88c2-27495348485e/kube-rbac-proxy/0.log" Oct 04 03:39:14 crc kubenswrapper[4742]: I1004 03:39:14.756430 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-fcc5d_0062ea87-e9b5-44f9-88c2-27495348485e/machine-api-operator/0.log" Oct 04 03:39:26 crc kubenswrapper[4742]: I1004 03:39:26.859443 4742 scope.go:117] "RemoveContainer" containerID="ad8cd41abb89c3ef4bb7f33e14da9242af2dc910209a036b049d8fc459c07508" Oct 04 03:39:26 crc kubenswrapper[4742]: E1004 03:39:26.860923 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qdt7v_openshift-machine-config-operator(989ea247-7ac9-4ade-8dcb-2632fef4f83e)\"" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" Oct 04 03:39:32 crc kubenswrapper[4742]: I1004 03:39:32.239384 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-cdc8r_9c0b8053-d605-4419-81fa-e7f17c45e068/kube-rbac-proxy/0.log" Oct 04 03:39:32 crc kubenswrapper[4742]: I1004 03:39:32.313687 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-cdc8r_9c0b8053-d605-4419-81fa-e7f17c45e068/controller/0.log" Oct 04 03:39:32 crc kubenswrapper[4742]: I1004 03:39:32.574119 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-b4g9q_c32d6520-f0f0-4ba4-8317-0208036377db/frr-k8s-webhook-server/0.log" Oct 04 03:39:32 crc kubenswrapper[4742]: I1004 03:39:32.653835 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zc8db_37f0f686-5a17-4d4a-87d8-0f1d3410530f/cp-frr-files/0.log" Oct 04 03:39:32 crc kubenswrapper[4742]: I1004 03:39:32.835864 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zc8db_37f0f686-5a17-4d4a-87d8-0f1d3410530f/cp-reloader/0.log" Oct 04 03:39:32 crc kubenswrapper[4742]: I1004 03:39:32.859490 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zc8db_37f0f686-5a17-4d4a-87d8-0f1d3410530f/cp-frr-files/0.log" Oct 04 03:39:32 crc kubenswrapper[4742]: I1004 03:39:32.873003 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zc8db_37f0f686-5a17-4d4a-87d8-0f1d3410530f/cp-reloader/0.log" Oct 04 03:39:32 crc kubenswrapper[4742]: I1004 03:39:32.895680 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zc8db_37f0f686-5a17-4d4a-87d8-0f1d3410530f/cp-metrics/0.log" Oct 04 03:39:33 crc kubenswrapper[4742]: I1004 03:39:33.082028 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zc8db_37f0f686-5a17-4d4a-87d8-0f1d3410530f/cp-frr-files/0.log" Oct 04 03:39:33 crc kubenswrapper[4742]: I1004 03:39:33.093098 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zc8db_37f0f686-5a17-4d4a-87d8-0f1d3410530f/cp-reloader/0.log" Oct 04 03:39:33 crc kubenswrapper[4742]: I1004 03:39:33.128005 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zc8db_37f0f686-5a17-4d4a-87d8-0f1d3410530f/cp-metrics/0.log" Oct 04 03:39:33 crc kubenswrapper[4742]: I1004 03:39:33.145082 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zc8db_37f0f686-5a17-4d4a-87d8-0f1d3410530f/cp-metrics/0.log" Oct 04 03:39:33 crc kubenswrapper[4742]: I1004 03:39:33.293537 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zc8db_37f0f686-5a17-4d4a-87d8-0f1d3410530f/cp-reloader/0.log" Oct 04 03:39:33 crc kubenswrapper[4742]: I1004 03:39:33.296187 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zc8db_37f0f686-5a17-4d4a-87d8-0f1d3410530f/cp-frr-files/0.log" Oct 04 03:39:33 crc kubenswrapper[4742]: I1004 03:39:33.321593 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zc8db_37f0f686-5a17-4d4a-87d8-0f1d3410530f/cp-metrics/0.log" Oct 04 03:39:33 crc kubenswrapper[4742]: I1004 03:39:33.356846 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zc8db_37f0f686-5a17-4d4a-87d8-0f1d3410530f/controller/0.log" Oct 04 03:39:33 crc kubenswrapper[4742]: I1004 03:39:33.542710 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zc8db_37f0f686-5a17-4d4a-87d8-0f1d3410530f/frr-metrics/0.log" Oct 04 03:39:33 crc kubenswrapper[4742]: I1004 03:39:33.556575 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zc8db_37f0f686-5a17-4d4a-87d8-0f1d3410530f/kube-rbac-proxy/0.log" Oct 04 03:39:33 crc kubenswrapper[4742]: I1004 03:39:33.597245 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zc8db_37f0f686-5a17-4d4a-87d8-0f1d3410530f/kube-rbac-proxy-frr/0.log" Oct 04 03:39:33 crc kubenswrapper[4742]: I1004 03:39:33.760815 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zc8db_37f0f686-5a17-4d4a-87d8-0f1d3410530f/reloader/0.log" Oct 04 03:39:33 crc kubenswrapper[4742]: I1004 03:39:33.840930 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-c97f8c589-xphrs_9bcd94cc-6df7-4f9c-be06-2234b0d99d72/manager/0.log" Oct 04 03:39:33 crc kubenswrapper[4742]: I1004 03:39:33.969522 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zc8db_37f0f686-5a17-4d4a-87d8-0f1d3410530f/frr/0.log" Oct 04 03:39:34 crc kubenswrapper[4742]: I1004 03:39:34.015931 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-75697b4fb9-tnl24_e401237a-d63b-4e6a-8efc-58b9ed5a6846/webhook-server/0.log" Oct 04 03:39:34 crc kubenswrapper[4742]: I1004 03:39:34.153551 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-r985t_b43429d7-ac32-4ab9-91e1-1ac6010312ec/kube-rbac-proxy/0.log" Oct 04 03:39:34 crc kubenswrapper[4742]: I1004 03:39:34.291247 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-r985t_b43429d7-ac32-4ab9-91e1-1ac6010312ec/speaker/0.log" Oct 04 03:39:41 crc kubenswrapper[4742]: I1004 03:39:41.858822 4742 scope.go:117] "RemoveContainer" containerID="ad8cd41abb89c3ef4bb7f33e14da9242af2dc910209a036b049d8fc459c07508" Oct 04 03:39:41 crc kubenswrapper[4742]: E1004 03:39:41.859674 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qdt7v_openshift-machine-config-operator(989ea247-7ac9-4ade-8dcb-2632fef4f83e)\"" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" Oct 04 03:39:47 crc kubenswrapper[4742]: I1004 03:39:47.045922 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-cm85w"] Oct 04 03:39:47 crc kubenswrapper[4742]: I1004 03:39:47.056605 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-cm85w"] Oct 04 03:39:48 crc kubenswrapper[4742]: I1004 03:39:48.869400 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a438233-cd6d-4fbe-8008-d079ebd72d20" path="/var/lib/kubelet/pods/8a438233-cd6d-4fbe-8008-d079ebd72d20/volumes" Oct 04 03:39:50 crc kubenswrapper[4742]: I1004 03:39:50.228856 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-0305-account-create-frmkd_fa270305-d910-44a7-956c-27b7268d448f/mariadb-account-create/0.log" Oct 04 03:39:50 crc kubenswrapper[4742]: I1004 03:39:50.406738 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-db-sync-mwk7z_679af79d-4438-49e7-ace8-d83ff23a2a97/glance-db-sync/0.log" Oct 04 03:39:50 crc kubenswrapper[4742]: I1004 03:39:50.502466 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-external-api-0_7cad2cc9-60ac-4e19-81cc-4676387a2176/glance-api/0.log" Oct 04 03:39:50 crc kubenswrapper[4742]: I1004 03:39:50.511607 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-external-api-0_7cad2cc9-60ac-4e19-81cc-4676387a2176/glance-httpd/0.log" Oct 04 03:39:50 crc kubenswrapper[4742]: I1004 03:39:50.624053 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-external-api-0_7cad2cc9-60ac-4e19-81cc-4676387a2176/glance-log/0.log" Oct 04 03:39:50 crc kubenswrapper[4742]: I1004 03:39:50.713929 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-internal-api-0_43cf2eff-93e3-4311-ac2f-ef3ecdfeb970/glance-api/0.log" Oct 04 03:39:50 crc kubenswrapper[4742]: I1004 03:39:50.718523 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-internal-api-0_43cf2eff-93e3-4311-ac2f-ef3ecdfeb970/glance-httpd/0.log" Oct 04 03:39:50 crc kubenswrapper[4742]: I1004 03:39:50.875336 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-internal-api-0_43cf2eff-93e3-4311-ac2f-ef3ecdfeb970/glance-log/0.log" Oct 04 03:39:51 crc kubenswrapper[4742]: I1004 03:39:51.207985 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_memcached-0_745e8f2b-0963-4aed-b04e-a21ddc5c9ac7/memcached/0.log" Oct 04 03:39:51 crc kubenswrapper[4742]: I1004 03:39:51.216075 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_keystone-5b74dd9df7-n4fhz_f99f948a-5fbd-4292-abcd-c8c64992634e/keystone-api/0.log" Oct 04 03:39:51 crc kubenswrapper[4742]: I1004 03:39:51.249568 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_1b69f4d7-789b-4992-9b33-a839e3105797/mysql-bootstrap/0.log" Oct 04 03:39:51 crc kubenswrapper[4742]: I1004 03:39:51.496493 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_1b69f4d7-789b-4992-9b33-a839e3105797/galera/0.log" Oct 04 03:39:51 crc kubenswrapper[4742]: I1004 03:39:51.513804 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_1b69f4d7-789b-4992-9b33-a839e3105797/mysql-bootstrap/0.log" Oct 04 03:39:51 crc kubenswrapper[4742]: I1004 03:39:51.557613 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_68095fb7-c650-4e91-aa84-1a66a90076db/mysql-bootstrap/0.log" Oct 04 03:39:51 crc kubenswrapper[4742]: I1004 03:39:51.727603 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_68095fb7-c650-4e91-aa84-1a66a90076db/galera/0.log" Oct 04 03:39:51 crc kubenswrapper[4742]: I1004 03:39:51.750504 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_68095fb7-c650-4e91-aa84-1a66a90076db/mysql-bootstrap/0.log" Oct 04 03:39:51 crc kubenswrapper[4742]: I1004 03:39:51.773962 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_4786b0c4-6846-4873-a3fb-0e155d15e2d0/mysql-bootstrap/0.log" Oct 04 03:39:51 crc kubenswrapper[4742]: I1004 03:39:51.966530 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_4786b0c4-6846-4873-a3fb-0e155d15e2d0/mysql-bootstrap/0.log" Oct 04 03:39:52 crc kubenswrapper[4742]: I1004 03:39:52.035977 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_4786b0c4-6846-4873-a3fb-0e155d15e2d0/galera/0.log" Oct 04 03:39:52 crc kubenswrapper[4742]: I1004 03:39:52.205022 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstackclient_1bc82217-289b-4cff-9f9a-1b62bbfd03f6/openstackclient/0.log" Oct 04 03:39:52 crc kubenswrapper[4742]: I1004 03:39:52.404854 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_8e6c595f-148d-419e-8a72-a991c62df809/setup-container/0.log" Oct 04 03:39:52 crc kubenswrapper[4742]: I1004 03:39:52.597808 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_8e6c595f-148d-419e-8a72-a991c62df809/setup-container/0.log" Oct 04 03:39:52 crc kubenswrapper[4742]: I1004 03:39:52.674625 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-proxy-59cb459c9f-hs28w_5d409f2c-d078-4589-87c8-7e6fb42e667b/proxy-httpd/0.log" Oct 04 03:39:52 crc kubenswrapper[4742]: I1004 03:39:52.684431 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_8e6c595f-148d-419e-8a72-a991c62df809/rabbitmq/0.log" Oct 04 03:39:52 crc kubenswrapper[4742]: I1004 03:39:52.831552 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-proxy-59cb459c9f-hs28w_5d409f2c-d078-4589-87c8-7e6fb42e667b/proxy-server/0.log" Oct 04 03:39:52 crc kubenswrapper[4742]: I1004 03:39:52.897591 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-ring-rebalance-jvgv2_70411213-953f-4207-a31a-4dd2214efd9e/swift-ring-rebalance/0.log" Oct 04 03:39:53 crc kubenswrapper[4742]: I1004 03:39:53.035594 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_7e8086e1-26ed-40f6-a3f7-861ec4d232f9/account-auditor/0.log" Oct 04 03:39:53 crc kubenswrapper[4742]: I1004 03:39:53.078404 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_7e8086e1-26ed-40f6-a3f7-861ec4d232f9/account-reaper/0.log" Oct 04 03:39:53 crc kubenswrapper[4742]: I1004 03:39:53.101578 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_7e8086e1-26ed-40f6-a3f7-861ec4d232f9/account-replicator/0.log" Oct 04 03:39:53 crc kubenswrapper[4742]: I1004 03:39:53.236372 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_7e8086e1-26ed-40f6-a3f7-861ec4d232f9/account-server/0.log" Oct 04 03:39:53 crc kubenswrapper[4742]: I1004 03:39:53.249098 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_7e8086e1-26ed-40f6-a3f7-861ec4d232f9/container-auditor/0.log" Oct 04 03:39:53 crc kubenswrapper[4742]: I1004 03:39:53.292955 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_7e8086e1-26ed-40f6-a3f7-861ec4d232f9/container-replicator/0.log" Oct 04 03:39:53 crc kubenswrapper[4742]: I1004 03:39:53.359470 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_7e8086e1-26ed-40f6-a3f7-861ec4d232f9/container-server/0.log" Oct 04 03:39:53 crc kubenswrapper[4742]: I1004 03:39:53.457036 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_7e8086e1-26ed-40f6-a3f7-861ec4d232f9/container-updater/0.log" Oct 04 03:39:53 crc kubenswrapper[4742]: I1004 03:39:53.485680 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_7e8086e1-26ed-40f6-a3f7-861ec4d232f9/object-auditor/0.log" Oct 04 03:39:53 crc kubenswrapper[4742]: I1004 03:39:53.509142 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_7e8086e1-26ed-40f6-a3f7-861ec4d232f9/object-expirer/0.log" Oct 04 03:39:53 crc kubenswrapper[4742]: I1004 03:39:53.518529 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_7e8086e1-26ed-40f6-a3f7-861ec4d232f9/object-replicator/0.log" Oct 04 03:39:53 crc kubenswrapper[4742]: I1004 03:39:53.559909 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_7e8086e1-26ed-40f6-a3f7-861ec4d232f9/object-server/0.log" Oct 04 03:39:53 crc kubenswrapper[4742]: I1004 03:39:53.689855 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_7e8086e1-26ed-40f6-a3f7-861ec4d232f9/rsync/0.log" Oct 04 03:39:53 crc kubenswrapper[4742]: I1004 03:39:53.699130 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_7e8086e1-26ed-40f6-a3f7-861ec4d232f9/object-updater/0.log" Oct 04 03:39:53 crc kubenswrapper[4742]: I1004 03:39:53.707793 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_7e8086e1-26ed-40f6-a3f7-861ec4d232f9/swift-recon-cron/0.log" Oct 04 03:39:55 crc kubenswrapper[4742]: I1004 03:39:55.858368 4742 scope.go:117] "RemoveContainer" containerID="ad8cd41abb89c3ef4bb7f33e14da9242af2dc910209a036b049d8fc459c07508" Oct 04 03:39:55 crc kubenswrapper[4742]: E1004 03:39:55.859233 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qdt7v_openshift-machine-config-operator(989ea247-7ac9-4ade-8dcb-2632fef4f83e)\"" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" Oct 04 03:39:58 crc kubenswrapper[4742]: I1004 03:39:58.041361 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-0305-account-create-frmkd"] Oct 04 03:39:58 crc kubenswrapper[4742]: I1004 03:39:58.047260 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-0305-account-create-frmkd"] Oct 04 03:39:58 crc kubenswrapper[4742]: I1004 03:39:58.871854 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa270305-d910-44a7-956c-27b7268d448f" path="/var/lib/kubelet/pods/fa270305-d910-44a7-956c-27b7268d448f/volumes" Oct 04 03:40:06 crc kubenswrapper[4742]: I1004 03:40:06.039949 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-mwk7z"] Oct 04 03:40:06 crc kubenswrapper[4742]: I1004 03:40:06.053020 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-mwk7z"] Oct 04 03:40:06 crc kubenswrapper[4742]: I1004 03:40:06.858839 4742 scope.go:117] "RemoveContainer" containerID="ad8cd41abb89c3ef4bb7f33e14da9242af2dc910209a036b049d8fc459c07508" Oct 04 03:40:06 crc kubenswrapper[4742]: E1004 03:40:06.859233 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qdt7v_openshift-machine-config-operator(989ea247-7ac9-4ade-8dcb-2632fef4f83e)\"" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" Oct 04 03:40:06 crc kubenswrapper[4742]: I1004 03:40:06.874992 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="679af79d-4438-49e7-ace8-d83ff23a2a97" path="/var/lib/kubelet/pods/679af79d-4438-49e7-ace8-d83ff23a2a97/volumes" Oct 04 03:40:08 crc kubenswrapper[4742]: I1004 03:40:08.971923 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6_5a2baed0-2265-4564-87fe-ca79418865f6/util/0.log" Oct 04 03:40:09 crc kubenswrapper[4742]: I1004 03:40:09.164441 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6_5a2baed0-2265-4564-87fe-ca79418865f6/pull/0.log" Oct 04 03:40:09 crc kubenswrapper[4742]: I1004 03:40:09.165175 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6_5a2baed0-2265-4564-87fe-ca79418865f6/pull/0.log" Oct 04 03:40:09 crc kubenswrapper[4742]: I1004 03:40:09.198257 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6_5a2baed0-2265-4564-87fe-ca79418865f6/util/0.log" Oct 04 03:40:09 crc kubenswrapper[4742]: I1004 03:40:09.353398 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6_5a2baed0-2265-4564-87fe-ca79418865f6/pull/0.log" Oct 04 03:40:09 crc kubenswrapper[4742]: I1004 03:40:09.382711 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6_5a2baed0-2265-4564-87fe-ca79418865f6/extract/0.log" Oct 04 03:40:09 crc kubenswrapper[4742]: I1004 03:40:09.408027 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d24s4c6_5a2baed0-2265-4564-87fe-ca79418865f6/util/0.log" Oct 04 03:40:09 crc kubenswrapper[4742]: I1004 03:40:09.541575 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-l5vff_3e50888f-9b21-423e-a450-66f235ad93c8/extract-utilities/0.log" Oct 04 03:40:09 crc kubenswrapper[4742]: I1004 03:40:09.718688 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-l5vff_3e50888f-9b21-423e-a450-66f235ad93c8/extract-utilities/0.log" Oct 04 03:40:09 crc kubenswrapper[4742]: I1004 03:40:09.746062 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-l5vff_3e50888f-9b21-423e-a450-66f235ad93c8/extract-content/0.log" Oct 04 03:40:09 crc kubenswrapper[4742]: I1004 03:40:09.769956 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-l5vff_3e50888f-9b21-423e-a450-66f235ad93c8/extract-content/0.log" Oct 04 03:40:09 crc kubenswrapper[4742]: I1004 03:40:09.929089 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-l5vff_3e50888f-9b21-423e-a450-66f235ad93c8/extract-content/0.log" Oct 04 03:40:09 crc kubenswrapper[4742]: I1004 03:40:09.991818 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-l5vff_3e50888f-9b21-423e-a450-66f235ad93c8/extract-utilities/0.log" Oct 04 03:40:10 crc kubenswrapper[4742]: I1004 03:40:10.203075 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t2zpx_4bde2e58-49b8-4975-aaaf-2cecd14c34d1/extract-utilities/0.log" Oct 04 03:40:10 crc kubenswrapper[4742]: I1004 03:40:10.399455 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t2zpx_4bde2e58-49b8-4975-aaaf-2cecd14c34d1/extract-content/0.log" Oct 04 03:40:10 crc kubenswrapper[4742]: I1004 03:40:10.431475 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-l5vff_3e50888f-9b21-423e-a450-66f235ad93c8/registry-server/0.log" Oct 04 03:40:10 crc kubenswrapper[4742]: I1004 03:40:10.498089 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t2zpx_4bde2e58-49b8-4975-aaaf-2cecd14c34d1/extract-utilities/0.log" Oct 04 03:40:10 crc kubenswrapper[4742]: I1004 03:40:10.508511 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t2zpx_4bde2e58-49b8-4975-aaaf-2cecd14c34d1/extract-content/0.log" Oct 04 03:40:10 crc kubenswrapper[4742]: I1004 03:40:10.615480 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t2zpx_4bde2e58-49b8-4975-aaaf-2cecd14c34d1/extract-utilities/0.log" Oct 04 03:40:10 crc kubenswrapper[4742]: I1004 03:40:10.653423 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t2zpx_4bde2e58-49b8-4975-aaaf-2cecd14c34d1/extract-content/0.log" Oct 04 03:40:10 crc kubenswrapper[4742]: I1004 03:40:10.829560 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-4mmhw_18ed4cd4-2848-4f6f-bf76-bd1425aa40b5/marketplace-operator/0.log" Oct 04 03:40:11 crc kubenswrapper[4742]: I1004 03:40:11.012544 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c29rn_1ac418c2-2977-4273-897b-282809b825d0/extract-utilities/0.log" Oct 04 03:40:11 crc kubenswrapper[4742]: I1004 03:40:11.068094 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t2zpx_4bde2e58-49b8-4975-aaaf-2cecd14c34d1/registry-server/0.log" Oct 04 03:40:11 crc kubenswrapper[4742]: I1004 03:40:11.259479 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c29rn_1ac418c2-2977-4273-897b-282809b825d0/extract-content/0.log" Oct 04 03:40:11 crc kubenswrapper[4742]: I1004 03:40:11.266453 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c29rn_1ac418c2-2977-4273-897b-282809b825d0/extract-content/0.log" Oct 04 03:40:11 crc kubenswrapper[4742]: I1004 03:40:11.279327 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c29rn_1ac418c2-2977-4273-897b-282809b825d0/extract-utilities/0.log" Oct 04 03:40:11 crc kubenswrapper[4742]: I1004 03:40:11.475418 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c29rn_1ac418c2-2977-4273-897b-282809b825d0/extract-utilities/0.log" Oct 04 03:40:11 crc kubenswrapper[4742]: I1004 03:40:11.500105 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c29rn_1ac418c2-2977-4273-897b-282809b825d0/extract-content/0.log" Oct 04 03:40:11 crc kubenswrapper[4742]: I1004 03:40:11.532010 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c29rn_1ac418c2-2977-4273-897b-282809b825d0/registry-server/0.log" Oct 04 03:40:11 crc kubenswrapper[4742]: I1004 03:40:11.661979 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9ztnx_f4f312f8-6a5a-41da-9de0-0e67ea5b3344/extract-utilities/0.log" Oct 04 03:40:11 crc kubenswrapper[4742]: I1004 03:40:11.832774 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9ztnx_f4f312f8-6a5a-41da-9de0-0e67ea5b3344/extract-utilities/0.log" Oct 04 03:40:11 crc kubenswrapper[4742]: I1004 03:40:11.851440 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9ztnx_f4f312f8-6a5a-41da-9de0-0e67ea5b3344/extract-content/0.log" Oct 04 03:40:11 crc kubenswrapper[4742]: I1004 03:40:11.885929 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9ztnx_f4f312f8-6a5a-41da-9de0-0e67ea5b3344/extract-content/0.log" Oct 04 03:40:12 crc kubenswrapper[4742]: I1004 03:40:12.031712 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9ztnx_f4f312f8-6a5a-41da-9de0-0e67ea5b3344/extract-utilities/0.log" Oct 04 03:40:12 crc kubenswrapper[4742]: I1004 03:40:12.035744 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9ztnx_f4f312f8-6a5a-41da-9de0-0e67ea5b3344/extract-content/0.log" Oct 04 03:40:12 crc kubenswrapper[4742]: I1004 03:40:12.524642 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9ztnx_f4f312f8-6a5a-41da-9de0-0e67ea5b3344/registry-server/0.log" Oct 04 03:40:18 crc kubenswrapper[4742]: I1004 03:40:18.858550 4742 scope.go:117] "RemoveContainer" containerID="ad8cd41abb89c3ef4bb7f33e14da9242af2dc910209a036b049d8fc459c07508" Oct 04 03:40:18 crc kubenswrapper[4742]: E1004 03:40:18.859491 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qdt7v_openshift-machine-config-operator(989ea247-7ac9-4ade-8dcb-2632fef4f83e)\"" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" Oct 04 03:40:30 crc kubenswrapper[4742]: I1004 03:40:30.858440 4742 scope.go:117] "RemoveContainer" containerID="ad8cd41abb89c3ef4bb7f33e14da9242af2dc910209a036b049d8fc459c07508" Oct 04 03:40:30 crc kubenswrapper[4742]: E1004 03:40:30.859464 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qdt7v_openshift-machine-config-operator(989ea247-7ac9-4ade-8dcb-2632fef4f83e)\"" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" Oct 04 03:40:36 crc kubenswrapper[4742]: I1004 03:40:36.424885 4742 scope.go:117] "RemoveContainer" containerID="8626a54ac197404e9c659c4d52753ead32c05a2d4db068d29eb3ad48882ee0fc" Oct 04 03:40:36 crc kubenswrapper[4742]: I1004 03:40:36.461954 4742 scope.go:117] "RemoveContainer" containerID="6e451cad86f2dfcf4fdafd7eda8e2daf1c97f2a8ebb5e09649553cdfc7c2ac3e" Oct 04 03:40:36 crc kubenswrapper[4742]: I1004 03:40:36.551426 4742 scope.go:117] "RemoveContainer" containerID="daf788729c878af0db8fa608f6798c069e3449d29172b3b63ad756cb4e0a105c" Oct 04 03:40:43 crc kubenswrapper[4742]: I1004 03:40:43.859510 4742 scope.go:117] "RemoveContainer" containerID="ad8cd41abb89c3ef4bb7f33e14da9242af2dc910209a036b049d8fc459c07508" Oct 04 03:40:43 crc kubenswrapper[4742]: E1004 03:40:43.860919 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qdt7v_openshift-machine-config-operator(989ea247-7ac9-4ade-8dcb-2632fef4f83e)\"" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" Oct 04 03:40:57 crc kubenswrapper[4742]: I1004 03:40:57.858757 4742 scope.go:117] "RemoveContainer" containerID="ad8cd41abb89c3ef4bb7f33e14da9242af2dc910209a036b049d8fc459c07508" Oct 04 03:40:57 crc kubenswrapper[4742]: E1004 03:40:57.859889 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qdt7v_openshift-machine-config-operator(989ea247-7ac9-4ade-8dcb-2632fef4f83e)\"" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" Oct 04 03:41:09 crc kubenswrapper[4742]: I1004 03:41:09.858604 4742 scope.go:117] "RemoveContainer" containerID="ad8cd41abb89c3ef4bb7f33e14da9242af2dc910209a036b049d8fc459c07508" Oct 04 03:41:09 crc kubenswrapper[4742]: E1004 03:41:09.859473 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qdt7v_openshift-machine-config-operator(989ea247-7ac9-4ade-8dcb-2632fef4f83e)\"" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" Oct 04 03:41:15 crc kubenswrapper[4742]: I1004 03:41:15.242532 4742 generic.go:334] "Generic (PLEG): container finished" podID="b6afdf41-80ea-433f-924c-2627f10d96a4" containerID="936b3262f0bd1f4c1bd43d8b35714763a5281cf1c26244629b85efd9e87d39f7" exitCode=0 Oct 04 03:41:15 crc kubenswrapper[4742]: I1004 03:41:15.242612 4742 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x49qd/must-gather-wcl4z" event={"ID":"b6afdf41-80ea-433f-924c-2627f10d96a4","Type":"ContainerDied","Data":"936b3262f0bd1f4c1bd43d8b35714763a5281cf1c26244629b85efd9e87d39f7"} Oct 04 03:41:15 crc kubenswrapper[4742]: I1004 03:41:15.243942 4742 scope.go:117] "RemoveContainer" containerID="936b3262f0bd1f4c1bd43d8b35714763a5281cf1c26244629b85efd9e87d39f7" Oct 04 03:41:16 crc kubenswrapper[4742]: I1004 03:41:16.209904 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-x49qd_must-gather-wcl4z_b6afdf41-80ea-433f-924c-2627f10d96a4/gather/0.log" Oct 04 03:41:21 crc kubenswrapper[4742]: I1004 03:41:21.858715 4742 scope.go:117] "RemoveContainer" containerID="ad8cd41abb89c3ef4bb7f33e14da9242af2dc910209a036b049d8fc459c07508" Oct 04 03:41:21 crc kubenswrapper[4742]: E1004 03:41:21.860234 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qdt7v_openshift-machine-config-operator(989ea247-7ac9-4ade-8dcb-2632fef4f83e)\"" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" Oct 04 03:41:23 crc kubenswrapper[4742]: I1004 03:41:23.454483 4742 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-x49qd/must-gather-wcl4z"] Oct 04 03:41:23 crc kubenswrapper[4742]: I1004 03:41:23.455176 4742 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-x49qd/must-gather-wcl4z" podUID="b6afdf41-80ea-433f-924c-2627f10d96a4" containerName="copy" containerID="cri-o://242dca1004b99a0f29ad156d3edfe785a3c6ae74c39d5c96900b8bdd4866ba5e" gracePeriod=2 Oct 04 03:41:23 crc kubenswrapper[4742]: I1004 03:41:23.460489 4742 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-x49qd/must-gather-wcl4z"] Oct 04 03:41:23 crc kubenswrapper[4742]: I1004 03:41:23.830830 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-x49qd_must-gather-wcl4z_b6afdf41-80ea-433f-924c-2627f10d96a4/copy/0.log" Oct 04 03:41:23 crc kubenswrapper[4742]: I1004 03:41:23.831523 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x49qd/must-gather-wcl4z" Oct 04 03:41:23 crc kubenswrapper[4742]: I1004 03:41:23.974606 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7h6kc\" (UniqueName: \"kubernetes.io/projected/b6afdf41-80ea-433f-924c-2627f10d96a4-kube-api-access-7h6kc\") pod \"b6afdf41-80ea-433f-924c-2627f10d96a4\" (UID: \"b6afdf41-80ea-433f-924c-2627f10d96a4\") " Oct 04 03:41:23 crc kubenswrapper[4742]: I1004 03:41:23.975003 4742 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b6afdf41-80ea-433f-924c-2627f10d96a4-must-gather-output\") pod \"b6afdf41-80ea-433f-924c-2627f10d96a4\" (UID: \"b6afdf41-80ea-433f-924c-2627f10d96a4\") " Oct 04 03:41:23 crc kubenswrapper[4742]: I1004 03:41:23.980232 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6afdf41-80ea-433f-924c-2627f10d96a4-kube-api-access-7h6kc" (OuterVolumeSpecName: "kube-api-access-7h6kc") pod "b6afdf41-80ea-433f-924c-2627f10d96a4" (UID: "b6afdf41-80ea-433f-924c-2627f10d96a4"). InnerVolumeSpecName "kube-api-access-7h6kc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:41:24 crc kubenswrapper[4742]: I1004 03:41:24.046814 4742 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6afdf41-80ea-433f-924c-2627f10d96a4-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "b6afdf41-80ea-433f-924c-2627f10d96a4" (UID: "b6afdf41-80ea-433f-924c-2627f10d96a4"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:41:24 crc kubenswrapper[4742]: I1004 03:41:24.076971 4742 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7h6kc\" (UniqueName: \"kubernetes.io/projected/b6afdf41-80ea-433f-924c-2627f10d96a4-kube-api-access-7h6kc\") on node \"crc\" DevicePath \"\"" Oct 04 03:41:24 crc kubenswrapper[4742]: I1004 03:41:24.077008 4742 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b6afdf41-80ea-433f-924c-2627f10d96a4-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 04 03:41:24 crc kubenswrapper[4742]: I1004 03:41:24.346992 4742 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-x49qd_must-gather-wcl4z_b6afdf41-80ea-433f-924c-2627f10d96a4/copy/0.log" Oct 04 03:41:24 crc kubenswrapper[4742]: I1004 03:41:24.347482 4742 generic.go:334] "Generic (PLEG): container finished" podID="b6afdf41-80ea-433f-924c-2627f10d96a4" containerID="242dca1004b99a0f29ad156d3edfe785a3c6ae74c39d5c96900b8bdd4866ba5e" exitCode=143 Oct 04 03:41:24 crc kubenswrapper[4742]: I1004 03:41:24.347534 4742 scope.go:117] "RemoveContainer" containerID="242dca1004b99a0f29ad156d3edfe785a3c6ae74c39d5c96900b8bdd4866ba5e" Oct 04 03:41:24 crc kubenswrapper[4742]: I1004 03:41:24.347643 4742 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x49qd/must-gather-wcl4z" Oct 04 03:41:24 crc kubenswrapper[4742]: I1004 03:41:24.377753 4742 scope.go:117] "RemoveContainer" containerID="936b3262f0bd1f4c1bd43d8b35714763a5281cf1c26244629b85efd9e87d39f7" Oct 04 03:41:24 crc kubenswrapper[4742]: I1004 03:41:24.438707 4742 scope.go:117] "RemoveContainer" containerID="242dca1004b99a0f29ad156d3edfe785a3c6ae74c39d5c96900b8bdd4866ba5e" Oct 04 03:41:24 crc kubenswrapper[4742]: E1004 03:41:24.439226 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"242dca1004b99a0f29ad156d3edfe785a3c6ae74c39d5c96900b8bdd4866ba5e\": container with ID starting with 242dca1004b99a0f29ad156d3edfe785a3c6ae74c39d5c96900b8bdd4866ba5e not found: ID does not exist" containerID="242dca1004b99a0f29ad156d3edfe785a3c6ae74c39d5c96900b8bdd4866ba5e" Oct 04 03:41:24 crc kubenswrapper[4742]: I1004 03:41:24.439289 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"242dca1004b99a0f29ad156d3edfe785a3c6ae74c39d5c96900b8bdd4866ba5e"} err="failed to get container status \"242dca1004b99a0f29ad156d3edfe785a3c6ae74c39d5c96900b8bdd4866ba5e\": rpc error: code = NotFound desc = could not find container \"242dca1004b99a0f29ad156d3edfe785a3c6ae74c39d5c96900b8bdd4866ba5e\": container with ID starting with 242dca1004b99a0f29ad156d3edfe785a3c6ae74c39d5c96900b8bdd4866ba5e not found: ID does not exist" Oct 04 03:41:24 crc kubenswrapper[4742]: I1004 03:41:24.439320 4742 scope.go:117] "RemoveContainer" containerID="936b3262f0bd1f4c1bd43d8b35714763a5281cf1c26244629b85efd9e87d39f7" Oct 04 03:41:24 crc kubenswrapper[4742]: E1004 03:41:24.439645 4742 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"936b3262f0bd1f4c1bd43d8b35714763a5281cf1c26244629b85efd9e87d39f7\": container with ID starting with 936b3262f0bd1f4c1bd43d8b35714763a5281cf1c26244629b85efd9e87d39f7 not found: ID does not exist" containerID="936b3262f0bd1f4c1bd43d8b35714763a5281cf1c26244629b85efd9e87d39f7" Oct 04 03:41:24 crc kubenswrapper[4742]: I1004 03:41:24.439668 4742 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"936b3262f0bd1f4c1bd43d8b35714763a5281cf1c26244629b85efd9e87d39f7"} err="failed to get container status \"936b3262f0bd1f4c1bd43d8b35714763a5281cf1c26244629b85efd9e87d39f7\": rpc error: code = NotFound desc = could not find container \"936b3262f0bd1f4c1bd43d8b35714763a5281cf1c26244629b85efd9e87d39f7\": container with ID starting with 936b3262f0bd1f4c1bd43d8b35714763a5281cf1c26244629b85efd9e87d39f7 not found: ID does not exist" Oct 04 03:41:24 crc kubenswrapper[4742]: I1004 03:41:24.869027 4742 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6afdf41-80ea-433f-924c-2627f10d96a4" path="/var/lib/kubelet/pods/b6afdf41-80ea-433f-924c-2627f10d96a4/volumes" Oct 04 03:41:34 crc kubenswrapper[4742]: I1004 03:41:34.864453 4742 scope.go:117] "RemoveContainer" containerID="ad8cd41abb89c3ef4bb7f33e14da9242af2dc910209a036b049d8fc459c07508" Oct 04 03:41:34 crc kubenswrapper[4742]: E1004 03:41:34.866738 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qdt7v_openshift-machine-config-operator(989ea247-7ac9-4ade-8dcb-2632fef4f83e)\"" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" Oct 04 03:41:36 crc kubenswrapper[4742]: I1004 03:41:36.640736 4742 scope.go:117] "RemoveContainer" containerID="09a3337c5c36901e05f462452e880263a61504e1b824954cf89f3d9a8a596e10" Oct 04 03:41:36 crc kubenswrapper[4742]: I1004 03:41:36.697423 4742 scope.go:117] "RemoveContainer" containerID="a1456191f4dd0301be4fc2f5941f334269f6c62f644f356187d8673519040020" Oct 04 03:41:36 crc kubenswrapper[4742]: I1004 03:41:36.766066 4742 scope.go:117] "RemoveContainer" containerID="66d0cd2c6ab84ac020a771b76f8ca6680c761e48ddd36eb00c94c08cf5e902e8" Oct 04 03:41:36 crc kubenswrapper[4742]: I1004 03:41:36.806931 4742 scope.go:117] "RemoveContainer" containerID="74271f031297e4294ff64669dae090cb4db5c191611c19e59653af0689cfd266" Oct 04 03:41:36 crc kubenswrapper[4742]: I1004 03:41:36.838652 4742 scope.go:117] "RemoveContainer" containerID="906eb3821a0f8cd0fc91bb8982b4d0a8ab867090b4a2f8950a82b802e85e2a35" Oct 04 03:41:36 crc kubenswrapper[4742]: I1004 03:41:36.873686 4742 scope.go:117] "RemoveContainer" containerID="7bbcf3655a6f68689244f8782bd960d514b6a06d0501c8acbbe4c068d08cb887" Oct 04 03:41:48 crc kubenswrapper[4742]: I1004 03:41:48.859049 4742 scope.go:117] "RemoveContainer" containerID="ad8cd41abb89c3ef4bb7f33e14da9242af2dc910209a036b049d8fc459c07508" Oct 04 03:41:48 crc kubenswrapper[4742]: E1004 03:41:48.860126 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qdt7v_openshift-machine-config-operator(989ea247-7ac9-4ade-8dcb-2632fef4f83e)\"" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" Oct 04 03:42:02 crc kubenswrapper[4742]: I1004 03:42:02.859634 4742 scope.go:117] "RemoveContainer" containerID="ad8cd41abb89c3ef4bb7f33e14da9242af2dc910209a036b049d8fc459c07508" Oct 04 03:42:02 crc kubenswrapper[4742]: E1004 03:42:02.860519 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qdt7v_openshift-machine-config-operator(989ea247-7ac9-4ade-8dcb-2632fef4f83e)\"" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" Oct 04 03:42:13 crc kubenswrapper[4742]: I1004 03:42:13.859083 4742 scope.go:117] "RemoveContainer" containerID="ad8cd41abb89c3ef4bb7f33e14da9242af2dc910209a036b049d8fc459c07508" Oct 04 03:42:13 crc kubenswrapper[4742]: E1004 03:42:13.861016 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qdt7v_openshift-machine-config-operator(989ea247-7ac9-4ade-8dcb-2632fef4f83e)\"" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" Oct 04 03:42:26 crc kubenswrapper[4742]: I1004 03:42:26.858416 4742 scope.go:117] "RemoveContainer" containerID="ad8cd41abb89c3ef4bb7f33e14da9242af2dc910209a036b049d8fc459c07508" Oct 04 03:42:26 crc kubenswrapper[4742]: E1004 03:42:26.859535 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qdt7v_openshift-machine-config-operator(989ea247-7ac9-4ade-8dcb-2632fef4f83e)\"" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" Oct 04 03:42:39 crc kubenswrapper[4742]: I1004 03:42:39.858990 4742 scope.go:117] "RemoveContainer" containerID="ad8cd41abb89c3ef4bb7f33e14da9242af2dc910209a036b049d8fc459c07508" Oct 04 03:42:39 crc kubenswrapper[4742]: E1004 03:42:39.859893 4742 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qdt7v_openshift-machine-config-operator(989ea247-7ac9-4ade-8dcb-2632fef4f83e)\"" pod="openshift-machine-config-operator/machine-config-daemon-qdt7v" podUID="989ea247-7ac9-4ade-8dcb-2632fef4f83e" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515070113676024453 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015070113677017371 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015070107551016506 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015070107551015456 5ustar corecore